[ 531.219987] env[65121]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=65121) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 531.220234] env[65121]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=65121) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 531.221270] env[65121]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=65121) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 531.221606] env[65121]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 531.336176] env[65121]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=65121) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 531.345911] env[65121]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=65121) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 531.390471] env[65121]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 531.949500] env[65121]: INFO nova.virt.driver [None req-2d2781a7-4ae3-4ce1-b447-c11e1a853208 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 532.021182] env[65121]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 532.021378] env[65121]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 532.021436] env[65121]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=65121) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 535.190475] env[65121]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-b9a50a8e-fb51-4692-9b7a-07df5f2f7e26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.207019] env[65121]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=65121) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 535.207191] env[65121]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-61dcea4d-d524-4a69-bf08-79660995e720 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.241930] env[65121]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 88c03. [ 535.242154] env[65121]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.221s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 535.242784] env[65121]: INFO nova.virt.vmwareapi.driver [None req-2d2781a7-4ae3-4ce1-b447-c11e1a853208 None None] VMware vCenter version: 7.0.3 [ 535.246228] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1492913-1647-4111-a15b-f4ac05e08ec8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.264690] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9923ef56-2512-479b-b490-5010ec229d68 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.271450] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6850bc-47c4-438f-b69d-da6ad4318dac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.278960] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f72d8b-24a4-4a95-bd04-cec6b742cc0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.292698] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda9693f-cb16-4c43-a945-7eb99ac6fd4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.299387] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a158688-228d-47a3-bf0a-e6dad582f893 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.329843] env[65121]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-7c3810ac-6b1a-43b0-b556-d5d5cef7bb74 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.335963] env[65121]: DEBUG nova.virt.vmwareapi.driver [None req-2d2781a7-4ae3-4ce1-b447-c11e1a853208 None None] Extension org.openstack.compute already exists. {{(pid=65121) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 535.338577] env[65121]: INFO nova.compute.provider_config [None req-2d2781a7-4ae3-4ce1-b447-c11e1a853208 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 535.842115] env[65121]: DEBUG nova.context [None req-2d2781a7-4ae3-4ce1-b447-c11e1a853208 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),7e996fde-d59f-4846-b44c-56b86ca89844(cell1) {{(pid=65121) load_cells /opt/stack/nova/nova/context.py:472}} [ 535.842390] env[65121]: INFO nova.utils [None req-2d2781a7-4ae3-4ce1-b447-c11e1a853208 None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 535.844388] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 535.844603] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 535.845298] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 535.845740] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Acquiring lock "7e996fde-d59f-4846-b44c-56b86ca89844" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 535.845916] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Lock "7e996fde-d59f-4846-b44c-56b86ca89844" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 535.846944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Lock "7e996fde-d59f-4846-b44c-56b86ca89844" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 535.867110] env[65121]: INFO dbcounter [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Registered counter for database nova_cell0 [ 535.876154] env[65121]: INFO dbcounter [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Registered counter for database nova_cell1 [ 535.879463] env[65121]: DEBUG oslo_db.sqlalchemy.engines [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65121) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 535.879839] env[65121]: DEBUG oslo_db.sqlalchemy.engines [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65121) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 535.885097] env[65121]: ERROR nova.db.main.api [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 535.885097] env[65121]: func(*args, **kwargs) [ 535.885097] env[65121]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 535.885097] env[65121]: self.work.run() [ 535.885097] env[65121]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 535.885097] env[65121]: result = self.fn(*self.args, **self.kwargs) [ 535.885097] env[65121]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 535.885097] env[65121]: return func(*args, **kwargs) [ 535.885097] env[65121]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 535.885097] env[65121]: result = fn(*args, **kwargs) [ 535.885097] env[65121]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 535.885097] env[65121]: return f(*args, **kwargs) [ 535.885097] env[65121]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 535.885097] env[65121]: return db.service_get_minimum_version(context, binaries) [ 535.885097] env[65121]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 535.885097] env[65121]: _check_db_access() [ 535.885097] env[65121]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 535.885097] env[65121]: stacktrace = ''.join(traceback.format_stack()) [ 535.885097] env[65121]: [ 535.885954] env[65121]: ERROR nova.db.main.api [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 535.885954] env[65121]: func(*args, **kwargs) [ 535.885954] env[65121]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 535.885954] env[65121]: self.work.run() [ 535.885954] env[65121]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 535.885954] env[65121]: result = self.fn(*self.args, **self.kwargs) [ 535.885954] env[65121]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 535.885954] env[65121]: return func(*args, **kwargs) [ 535.885954] env[65121]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 535.885954] env[65121]: result = fn(*args, **kwargs) [ 535.885954] env[65121]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 535.885954] env[65121]: return f(*args, **kwargs) [ 535.885954] env[65121]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 535.885954] env[65121]: return db.service_get_minimum_version(context, binaries) [ 535.885954] env[65121]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 535.885954] env[65121]: _check_db_access() [ 535.885954] env[65121]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 535.885954] env[65121]: stacktrace = ''.join(traceback.format_stack()) [ 535.885954] env[65121]: [ 535.886445] env[65121]: WARNING nova.objects.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 535.886547] env[65121]: WARNING nova.objects.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Failed to get minimum service version for cell 7e996fde-d59f-4846-b44c-56b86ca89844 [ 535.886975] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Acquiring lock "singleton_lock" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.887151] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Acquired lock "singleton_lock" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 535.887374] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Releasing lock "singleton_lock" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 535.887680] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Full set of CONF: {{(pid=65121) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/service.py:275}} [ 535.887801] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ******************************************************************************** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 535.887912] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] Configuration options gathered from: {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 535.888045] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 535.888228] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 535.888343] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ================================================================================ {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 535.888533] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] allow_resize_to_same_host = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.888686] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] arq_binding_timeout = 300 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.888803] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] backdoor_port = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.888916] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] backdoor_socket = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.889092] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] block_device_allocate_retries = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.889259] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] block_device_allocate_retries_interval = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.889407] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cell_worker_thread_pool_size = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.889563] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cert = self.pem {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.889713] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.889865] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute_monitors = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.890022] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] config_dir = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.890179] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] config_drive_format = iso9660 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.890306] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.890456] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] config_source = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.890608] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] console_host = devstack {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.890755] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] control_exchange = nova {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.890897] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cpu_allocation_ratio = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.891063] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] daemon = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.891224] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] debug = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.891370] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] default_access_ip_network_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.891519] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] default_availability_zone = nova {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.891658] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] default_ephemeral_format = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.891803] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] default_green_pool_size = 1000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.892048] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.892215] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] default_schedule_zone = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.892366] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] default_thread_pool_size = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.892507] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] disk_allocation_ratio = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.892649] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] enable_new_services = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.892790] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] flat_injected = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.892930] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] force_config_drive = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.893137] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] force_raw_images = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.893305] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] graceful_shutdown_timeout = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.893453] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] heal_instance_info_cache_interval = -1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.893657] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] host = cpu-1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.893819] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.893967] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] initial_disk_allocation_ratio = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.894134] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] initial_ram_allocation_ratio = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.894357] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.894516] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instance_build_timeout = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.894663] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instance_delete_interval = 300 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.894818] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instance_format = [instance: %(uuid)s] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.894968] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instance_name_template = instance-%08x {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.895132] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instance_usage_audit = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.895288] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instance_usage_audit_period = month {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.895440] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.895591] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] instances_path = /opt/stack/data/nova/instances {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.895739] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] internal_service_availability_zone = internal {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.895877] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] key = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.896029] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] live_migration_retry_count = 30 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.896189] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_color = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.896335] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_config_append = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.896484] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.896628] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_dir = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.896770] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_file = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.896884] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_options = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897041] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_rotate_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897197] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_rotate_interval_type = days {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897347] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] log_rotation_type = none {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897463] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897575] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897725] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897874] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.897988] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.898153] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] long_rpc_timeout = 1800 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.898296] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] max_concurrent_builds = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.898437] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] max_concurrent_live_migrations = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.898576] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] max_concurrent_snapshots = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.898714] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] max_local_block_devices = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.898884] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] max_logfile_count = 30 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.899065] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] max_logfile_size_mb = 200 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.899220] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] maximum_instance_delete_attempts = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.899363] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] migrate_max_retries = -1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.899513] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] mkisofs_cmd = genisoimage {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.899703] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] my_block_storage_ip = 10.180.1.21 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.899822] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] my_ip = 10.180.1.21 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.900014] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.900172] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] network_allocate_retries = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.900334] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.900486] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] osapi_compute_unique_server_name_scope = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.900633] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] password_length = 12 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.900776] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] periodic_enable = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.900918] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] periodic_fuzzy_delay = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.901097] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] pointer_model = usbtablet {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.901270] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] preallocate_images = none {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.901417] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] publish_errors = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.901533] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] pybasedir = /opt/stack/nova {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.901672] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ram_allocation_ratio = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.901815] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] rate_limit_burst = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.901971] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] rate_limit_except_level = CRITICAL {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.902161] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] rate_limit_interval = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.902307] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] reboot_timeout = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.902451] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] reclaim_instance_interval = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.902592] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] record = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.902744] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] reimage_timeout_per_gb = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.902894] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] report_interval = 120 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.903050] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] rescue_timeout = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.903202] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] reserved_host_cpus = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.903346] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] reserved_host_disk_mb = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.903490] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] reserved_host_memory_mb = 512 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.903633] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] reserved_huge_pages = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.903775] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] resize_confirm_window = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.903917] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] resize_fs_using_block_device = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.904070] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] resume_guests_state_on_host_boot = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.904226] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.904393] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] rpc_response_timeout = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.904544] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] run_external_periodic_tasks = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.904695] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] running_deleted_instance_action = reap {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.904836] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] running_deleted_instance_poll_interval = 1800 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.904978] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] running_deleted_instance_timeout = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.905136] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler_instance_sync_interval = 120 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.905288] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_down_time = 720 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.905438] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] servicegroup_driver = db {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.905576] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] shell_completion = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.905717] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] shelved_offload_time = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.905857] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] shelved_poll_interval = 3600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.906014] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] shutdown_timeout = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.906173] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] source_is_ipv6 = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.906317] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ssl_only = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.906552] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.906706] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] sync_power_state_interval = 600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.906854] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] sync_power_state_pool_size = 1000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.907013] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] syslog_log_facility = LOG_USER {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.907170] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] tempdir = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.907316] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] thread_pool_statistic_period = -1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.907460] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] timeout_nbd = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.907613] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] transport_url = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.907759] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] update_resources_interval = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.907902] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] use_cow_images = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.908053] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] use_journal = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.908203] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] use_json = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.908345] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] use_rootwrap_daemon = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.908484] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] use_stderr = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.908626] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] use_syslog = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.908769] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vcpu_pin_set = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.908919] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plugging_is_fatal = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.909080] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plugging_timeout = 300 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.909234] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] virt_mkfs = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.909378] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] volume_usage_poll_interval = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.909520] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] watch_log_file = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.909673] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] web = /usr/share/spice-html5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 535.909842] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_concurrency.disable_process_locking = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.910425] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.910614] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_brick.lock_path = /opt/stack/data/os_brick {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.910777] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.910960] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.911159] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.911317] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.911477] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.911635] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.911788] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.911946] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.compute_link_prefix = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.912154] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.912322] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.dhcp_domain = novalocal {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.912475] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.enable_instance_password = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.912623] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.glance_link_prefix = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.912774] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.912931] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.913126] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.instance_list_per_project_cells = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.913286] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.list_records_by_skipping_down_cells = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.913437] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.local_metadata_per_cell = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.913590] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.max_limit = 1000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.913740] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.metadata_cache_expiration = 15 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.913902] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.neutron_default_project_id = default {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.914071] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.response_validation = warn {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.914227] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.use_neutron_default_nets = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.914399] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.914556] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.914707] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.914863] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.915027] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.vendordata_dynamic_targets = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.915184] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.vendordata_jsonfile_path = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.915351] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.915523] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.backend = dogpile.cache.memcached {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.915680] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.backend_argument = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.915827] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.backend_expiration_time = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.915982] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.config_prefix = cache.oslo {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.916148] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.debug_cache_backend = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.916296] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.enable_retry_client = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.916444] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.enable_socket_keepalive = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.916598] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.enabled = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.916749] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.enforce_fips_mode = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.916900] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.expiration_time = 600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.917079] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.hashclient_dead_timeout = 60.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.917234] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.hashclient_retry_attempts = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.917389] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.hashclient_retry_timeout = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.917540] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_dead_retry = 300 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.917685] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_password = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.917832] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.917979] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.918142] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_pool_maxsize = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.918289] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.918436] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_sasl_enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.918598] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.918751] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_socket_timeout = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.918896] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.memcache_username = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.919058] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.proxies = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.919215] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.redis_db = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.919361] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.redis_password = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.919519] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.redis_sentinel_service_name = mymaster {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.919681] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.919835] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.redis_server = localhost:6379 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.919986] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.redis_socket_timeout = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.920179] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.redis_username = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.920344] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.retry_attempts = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.920498] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.retry_delay = 0.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.920694] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.socket_keepalive_count = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.920859] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.socket_keepalive_idle = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.921014] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.socket_keepalive_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.921171] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.tls_allowed_ciphers = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.921314] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.tls_cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.921457] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.tls_certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.921601] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.tls_enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.921743] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cache.tls_keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.921898] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.922094] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.auth_type = password {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.922257] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.922420] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.catalog_info = volumev3::publicURL {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.922569] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.922717] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.922863] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.cross_az_attach = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.923056] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.debug = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.923223] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.endpoint_template = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.923375] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.http_retries = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.923532] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.923682] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.923841] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.os_region_name = RegionOne {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.923990] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.924163] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cinder.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.924342] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.924505] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.cpu_dedicated_set = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.924654] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.cpu_shared_set = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.924807] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.image_type_exclude_list = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.924951] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.925147] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.max_concurrent_disk_ops = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.925309] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.max_disk_devices_to_attach = -1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.925457] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.925610] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.925757] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.resource_provider_association_refresh = 300 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.925902] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.926061] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.shutdown_retry_interval = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.926232] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.926397] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] conductor.workers = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.926559] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] console.allowed_origins = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.926705] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] console.ssl_ciphers = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.926859] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] console.ssl_minimum_version = default {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.927022] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] consoleauth.enforce_session_timeout = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.927182] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] consoleauth.token_ttl = 600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.927341] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.927484] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.927631] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.927780] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.927924] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.928081] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.928235] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.928378] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.928521] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.928662] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.928805] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.928949] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.929107] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.929266] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.service_type = accelerator {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.929414] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.929557] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.929699] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.929842] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.930014] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.930235] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] cyborg.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.930417] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.asyncio_connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.930595] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.asyncio_slave_connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.930759] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.backend = sqlalchemy {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.930917] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.931087] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.connection_debug = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.931250] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.connection_parameters = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.931400] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.connection_recycle_time = 3600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.931551] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.connection_trace = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.931701] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.db_inc_retry_interval = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.931850] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.db_max_retries = 20 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.932070] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.db_max_retry_interval = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.932186] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.db_retry_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.932359] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.max_overflow = 50 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.932514] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.max_pool_size = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.932663] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.max_retries = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.932821] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.933017] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.mysql_wsrep_sync_wait = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.933127] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.pool_timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.933279] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.retry_interval = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.933419] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.slave_connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.933592] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.sqlite_synchronous = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.933745] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] database.use_db_reconnect = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.933901] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.asyncio_connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.934062] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.asyncio_slave_connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.934227] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.backend = sqlalchemy {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.934410] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.934573] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.connection_debug = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.934730] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.connection_parameters = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.934879] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.connection_recycle_time = 3600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.935067] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.connection_trace = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.935244] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.db_inc_retry_interval = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.935413] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.db_max_retries = 20 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.935575] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.db_max_retry_interval = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.935768] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.db_retry_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.936021] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.max_overflow = 50 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.936260] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.max_pool_size = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.936508] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.max_retries = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.936779] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.937053] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.937321] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.pool_timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.937585] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.retry_interval = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.937840] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.slave_connection = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.938106] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] api_database.sqlite_synchronous = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.938340] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] devices.enabled_mdev_types = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.938522] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.938688] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ephemeral_storage_encryption.default_format = luks {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.938841] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ephemeral_storage_encryption.enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.938998] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.939178] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.api_servers = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.939358] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.939514] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.939665] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.939810] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.939962] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.940120] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.debug = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.940279] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.default_trusted_certificate_ids = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.940427] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.enable_certificate_validation = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.940624] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.enable_rbd_download = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.940791] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.940947] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.941111] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.941262] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.941405] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.941553] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.num_retries = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.941705] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.rbd_ceph_conf = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.941851] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.rbd_connect_timeout = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.942037] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.rbd_pool = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.942207] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.rbd_user = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.942357] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.942502] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.942644] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.942797] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.service_type = image {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.942958] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.943149] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.943312] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.943458] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.943646] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.943844] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.verify_glance_signatures = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.944016] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] glance.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.944171] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] guestfs.debug = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.944330] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.944482] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.auth_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.944628] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.944774] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.944923] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.945082] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.945235] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.945380] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.945528] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.945672] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.945817] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.945961] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.946118] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.946266] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.946407] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.946559] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.service_type = shared-file-system {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.946710] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.share_apply_policy_timeout = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.946857] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.947008] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.947163] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.947341] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.947522] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.947753] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] manila.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.947935] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] mks.enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.948311] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.948496] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] image_cache.manager_interval = 2400 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.948652] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] image_cache.precache_concurrency = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.948809] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] image_cache.remove_unused_base_images = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.948965] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.949138] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.949305] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] image_cache.subdirectory_name = _base {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.949469] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.api_max_retries = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.949622] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.api_retry_interval = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.949772] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.949924] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.auth_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.950094] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.950246] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.950398] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.950546] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.conductor_group = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.950692] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.950837] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.950984] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.951156] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.951304] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.951449] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.951591] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.951741] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.peer_list = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.951887] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.952072] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.952237] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.serial_console_state_timeout = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.952387] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.952544] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.service_type = baremetal {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.952691] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.shard = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.952840] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.952990] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.953163] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.953308] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.953477] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.953626] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ironic.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.953795] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.953956] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] key_manager.fixed_key = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.954144] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.954295] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.barbican_api_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.954459] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.barbican_endpoint = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.954620] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.barbican_endpoint_type = public {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.954766] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.barbican_region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.954913] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.955071] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.955225] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.955371] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.955516] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.955664] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.number_of_retries = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.955809] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.retry_delay = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.955955] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.send_service_user_token = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.956116] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.956262] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.956407] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.verify_ssl = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.956547] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican.verify_ssl_path = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.956700] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.956862] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.auth_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.957020] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.957168] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.957317] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.957462] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.957603] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.957767] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.957915] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] barbican_service_user.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.958080] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.approle_role_id = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.958231] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.approle_secret_id = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.958389] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.kv_mountpoint = secret {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.958537] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.kv_path = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.958686] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.kv_version = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.958832] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.namespace = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.958979] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.root_token_id = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.959140] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.ssl_ca_crt_file = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.959297] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.timeout = 60.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.959446] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.use_ssl = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.959602] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.959762] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.959910] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.960091] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.960245] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.960390] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.960531] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.960676] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.960819] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.960960] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.961116] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.961264] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.961410] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.961561] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.961715] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.service_type = identity {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.961863] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.962039] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.962205] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.962351] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.962518] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.962664] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] keystone.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.962839] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.ceph_mount_options = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.963181] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.963345] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.connection_uri = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.963493] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.cpu_mode = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.963642] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.cpu_model_extra_flags = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.963797] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.cpu_models = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.963953] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.cpu_power_governor_high = performance {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.964122] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.cpu_power_governor_low = powersave {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.964274] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.cpu_power_management = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.964466] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.964624] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.device_detach_attempts = 8 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.964775] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.device_detach_timeout = 20 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.964924] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.disk_cachemodes = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.965081] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.disk_prefix = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.965235] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.enabled_perf_events = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.965389] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.file_backed_memory = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.965540] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.gid_maps = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.965689] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.hw_disk_discard = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.965833] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.hw_machine_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.965988] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.images_rbd_ceph_conf = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.966151] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.966300] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.966453] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.images_rbd_glance_store_name = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.966606] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.images_rbd_pool = rbd {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.966763] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.images_type = default {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.966907] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.images_volume_group = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.967065] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.inject_key = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.967218] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.inject_partition = -2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.967363] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.inject_password = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.967513] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.iscsi_iface = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.967660] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.iser_use_multipath = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.967809] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_bandwidth = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.967970] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.968136] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_downtime = 500 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.968288] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.968437] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.968584] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_inbound_addr = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.968731] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_parallel_connections = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.968878] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.969041] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_permit_post_copy = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.969198] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_scheme = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.969356] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_timeout_action = abort {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.969505] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_tunnelled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.969651] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_uri = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.969798] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.live_migration_with_native_tls = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.969966] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.max_queues = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.970139] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.970372] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.970522] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.nfs_mount_options = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.970806] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.970968] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.971133] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.num_iser_scan_tries = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.971284] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.num_memory_encrypted_guests = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.971434] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.971583] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.num_pcie_ports = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.971735] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.num_volume_scan_tries = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.971887] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.pmem_namespaces = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.972067] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.quobyte_client_cfg = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.972358] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.972529] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rbd_connect_timeout = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.972682] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.972829] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.972976] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rbd_secret_uuid = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.973139] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rbd_user = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.973291] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.973508] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.remote_filesystem_transport = ssh {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.973582] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rescue_image_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.973722] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rescue_kernel_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.973862] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rescue_ramdisk_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.974020] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.974166] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.rx_queue_size = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.974340] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.smbfs_mount_options = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.974631] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.974795] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.snapshot_compression = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.974943] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.snapshot_image_format = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.975176] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.975326] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.sparse_logical_volumes = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.975490] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.supported_tpm_secret_security = ['user', 'host'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.975640] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.swtpm_enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.975794] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.swtpm_group = tss {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.975947] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.swtpm_user = tss {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.976116] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.sysinfo_serial = unique {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.976267] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.tb_cache_size = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.976412] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.tx_queue_size = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.976564] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.uid_maps = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.976712] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.use_default_aio_mode_for_volumes = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.976858] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.use_virtio_for_bridges = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.977024] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.virt_type = kvm {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.977187] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.volume_clear = zero {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.977337] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.volume_clear_size = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.977484] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.volume_enforce_multipath = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.977633] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.volume_use_multipath = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.977778] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.vzstorage_cache_path = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.977929] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.978094] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.vzstorage_mount_group = qemu {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.978250] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.vzstorage_mount_opts = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.978403] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.978671] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.978836] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.vzstorage_mount_user = stack {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.978987] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.979166] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.979323] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.auth_type = password {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.979470] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.979613] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.979757] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.979908] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.980067] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.980225] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.default_floating_pool = public {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.980368] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.980515] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.extension_sync_interval = 600 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.980656] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.http_retries = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.980798] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.980939] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.981092] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.981249] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.981391] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.981545] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.ovs_bridge = br-int {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.981691] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.physnets = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.981845] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.region_name = RegionOne {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.982010] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.982206] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.service_metadata_proxy = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.982364] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.982522] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.service_type = network {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.982671] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.982815] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.982958] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.983117] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.983287] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.983437] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] neutron.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.983591] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] notifications.bdms_in_notifications = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.983753] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] notifications.default_level = INFO {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.983900] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] notifications.include_share_mapping = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.984071] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] notifications.notification_format = unversioned {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.984226] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] notifications.notify_on_state_change = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.984409] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.984581] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] pci.alias = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.984735] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] pci.device_spec = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.984884] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] pci.report_in_placement = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.985063] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.985291] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.auth_type = password {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.985477] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.985631] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.985775] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.985922] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.986079] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.986227] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.986371] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.default_domain_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.986509] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.default_domain_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.986649] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.domain_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.986786] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.domain_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.986925] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.987098] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.987250] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.987390] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.987528] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.987678] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.password = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.987819] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.project_domain_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.987966] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.project_domain_name = Default {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.988129] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.project_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.988315] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.project_name = service {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.988472] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.region_name = RegionOne {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.988615] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.988757] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.988909] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.service_type = placement {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.989068] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.989216] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.989358] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.989496] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.system_scope = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.989634] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.989773] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.trust_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.989912] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.user_domain_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.990082] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.user_domain_name = Default {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.990229] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.user_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.990387] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.username = nova {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.990548] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.990690] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] placement.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.990860] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.cores = 20 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.991016] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.count_usage_from_placement = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.991188] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.991358] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.injected_file_content_bytes = 10240 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.991510] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.injected_file_path_length = 255 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.991661] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.injected_files = 5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.991810] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.instances = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.991958] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.key_pairs = 100 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.992151] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.metadata_items = 128 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.992309] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.ram = 51200 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.992460] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.recheck_quota = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.992609] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.server_group_members = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.992757] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.server_groups = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.992945] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.993129] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] quota.unified_limits_resource_strategy = require {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.993290] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.993440] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.993587] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.image_metadata_prefilter = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.993734] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.993881] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.max_attempts = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.994041] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.max_placement_results = 1000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.994199] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.994382] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.query_placement_for_image_type_support = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.994538] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.994698] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] scheduler.workers = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.994855] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.995014] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.995183] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.995342] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.995494] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.995640] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.995787] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.995958] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.996133] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.host_subset_size = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.996284] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.996428] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.996574] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.996721] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.996863] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.997028] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.isolated_hosts = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.997189] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.isolated_images = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.997365] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.997514] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.997660] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.997803] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.pci_in_placement = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.997945] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.998103] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.998252] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.998392] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.998533] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.998676] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.998818] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.track_instance_changes = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.998972] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.999136] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] metrics.required = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.999280] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] metrics.weight_multiplier = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.999425] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.999569] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] metrics.weight_setting = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 535.999867] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.000034] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] serial_console.enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.000210] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] serial_console.port_range = 10000:20000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.000380] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.000533] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.000689] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] serial_console.serialproxy_port = 6083 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.000845] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.001021] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.auth_type = password {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.001168] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.001313] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.001461] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.001603] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.001744] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.001896] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.send_service_user_token = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.002078] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.002236] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] service_user.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.002394] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.agent_enabled = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.002541] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.002834] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.003032] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.003220] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.html5proxy_port = 6082 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.003374] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.image_compression = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.003519] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.jpeg_compression = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.003660] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.playback_compression = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.003804] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.require_secure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.003955] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.server_listen = 127.0.0.1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.004121] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.004411] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.004572] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.streaming_mode = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.004719] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] spice.zlib_compression = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.004872] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] upgrade_levels.baseapi = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.005044] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] upgrade_levels.compute = auto {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.005199] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] upgrade_levels.conductor = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.005342] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] upgrade_levels.scheduler = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.005496] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.005647] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.auth_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.005790] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.005935] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.006097] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.006274] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.006422] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.006571] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.006713] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vendordata_dynamic_auth.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.006871] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.api_retry_count = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.007028] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.ca_file = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.007195] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.cache_prefix = devstack-image-cache {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.007349] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.cluster_name = testcl1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.007497] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.connection_pool_size = 10 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.007641] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.console_delay_seconds = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.007794] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.datastore_regex = ^datastore.* {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.007992] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.008183] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.host_password = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.008338] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.host_port = 443 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.008496] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.host_username = administrator@vsphere.local {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.008648] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.insecure = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.008795] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.integration_bridge = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.008943] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.maximum_objects = 100 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.009110] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.pbm_default_policy = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.009286] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.pbm_enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.009434] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.pbm_wsdl_location = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.009586] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.009729] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.serial_port_proxy_uri = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.009900] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.serial_port_service_uri = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.010038] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.task_poll_interval = 0.5 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.010206] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.use_linked_clone = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.010362] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.vnc_keymap = en-us {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.010513] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.vnc_port = 5900 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.010661] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vmware.vnc_port_total = 10000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.010832] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.auth_schemes = ['none'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.010993] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.011279] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.011450] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.011607] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.novncproxy_port = 6080 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.011792] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.server_listen = 127.0.0.1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.011959] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.012159] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.vencrypt_ca_certs = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.012331] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.vencrypt_client_cert = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.012480] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vnc.vencrypt_client_key = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.012710] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.012890] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.disable_deep_image_inspection = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.013054] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.013207] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.013355] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.013504] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.disable_rootwrap = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.013648] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.enable_numa_live_migration = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.013793] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.013939] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.014106] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.014253] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.libvirt_disable_apic = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.014424] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.014579] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.014725] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.014869] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.015022] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.015177] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.015348] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.015496] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.015639] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.015789] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.015955] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.016120] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] wsgi.secure_proxy_ssl_header = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.016275] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] zvm.ca_file = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.016419] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] zvm.cloud_connector_url = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.016712] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.016874] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] zvm.reachable_timeout = 300 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.017043] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.017210] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.017370] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.connection_string = messaging:// {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.017523] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.enabled = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.017673] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.es_doc_type = notification {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.017819] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.es_scroll_size = 10000 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.017970] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.es_scroll_time = 2m {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.018155] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.filter_error_trace = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.018332] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.hmac_keys = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.018503] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.sentinel_service_name = mymaster {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.018668] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.socket_timeout = 0.1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.018811] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.trace_requests = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.018955] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler.trace_sqlalchemy = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.019153] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler_jaeger.process_tags = {} {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.019301] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler_jaeger.service_name_prefix = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.019449] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] profiler_otlp.service_name_prefix = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.019610] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.019761] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.019904] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.020061] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.020212] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.020355] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.020498] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.020644] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.020788] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.020940] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.021096] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.021286] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.021445] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.021594] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.021742] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.021893] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.022074] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.022234] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.022388] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.022537] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.022685] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.022834] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.022989] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.023146] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.023295] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.023439] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.023583] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.023727] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.023870] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.024025] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.024178] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.ssl = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.024377] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.024542] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.024691] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.024842] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.024992] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.ssl_version = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.025155] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.025323] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.025473] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_notifications.retry = -1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.025632] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.025787] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_messaging_notifications.transport_url = **** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.025943] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.auth_section = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.026104] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.auth_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.026252] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.cafile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.026393] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.certfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.026536] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.collect_timing = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.026676] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.connect_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.026816] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.connect_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.026955] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.endpoint_id = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.027231] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.endpoint_interface = publicURL {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.027303] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.endpoint_override = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.027431] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.endpoint_region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.027572] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.endpoint_service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.027713] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.endpoint_service_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.027856] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.insecure = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.027996] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.keyfile = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.028153] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.max_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.028293] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.min_version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.028431] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.region_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.028571] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.retriable_status_codes = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.028713] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.service_name = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.028851] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.service_type = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.028994] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.split_loggers = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.029150] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.status_code_retries = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.029290] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.status_code_retry_delay = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.029431] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.timeout = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.029573] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.valid_interfaces = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.029712] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_limit.version = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.029856] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_reports.file_event_handler = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.030012] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.030166] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] oslo_reports.log_dir = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.030356] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.030503] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.030646] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.030790] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.030938] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.031096] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.031244] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.031397] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.031541] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_ovs_privileged.group = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.031714] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.031874] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.032057] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.032294] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.032371] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] vif_plug_ovs_privileged.user = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.032526] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.flat_interface = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.032690] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.032849] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.033010] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.033227] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.033397] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.033556] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.033703] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.033867] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.034073] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_ovs.isolate_vif = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.034238] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.034419] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.034588] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.034797] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_ovs.ovsdb_interface = native {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.034947] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] os_vif_ovs.per_port_bridge = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.035163] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] privsep_osbrick.capabilities = [21, 2] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.035349] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] privsep_osbrick.group = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.035502] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] privsep_osbrick.helper_command = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.035650] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] privsep_osbrick.log_daemon_traceback = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.035799] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.035948] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.036117] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] privsep_osbrick.user = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.036361] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.036448] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] nova_sys_admin.group = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.036590] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] nova_sys_admin.helper_command = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.036733] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] nova_sys_admin.log_daemon_traceback = False {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.036912] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.037159] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.037328] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] nova_sys_admin.user = None {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 536.037454] env[65121]: DEBUG oslo_service.backend._eventlet.service [None req-4ff2b79d-6e19-4122-a2aa-13165922a484 None None] ******************************************************************************** {{(pid=65121) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 536.037970] env[65121]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 536.542393] env[65121]: WARNING nova.virt.vmwareapi.driver [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 536.543171] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Getting list of instances from cluster (obj){ [ 536.543171] env[65121]: value = "domain-c8" [ 536.543171] env[65121]: _type = "ClusterComputeResource" [ 536.543171] env[65121]: } {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 536.544275] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5272f1-dcb8-47fb-8fb6-9526e8ede1ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.553223] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Got total of 0 instances {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 536.553689] env[65121]: INFO nova.virt.node [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Generated node identity d56783bf-3ede-475a-8c5a-8d8303049e47 [ 536.553923] env[65121]: INFO nova.virt.node [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Wrote node identity d56783bf-3ede-475a-8c5a-8d8303049e47 to /opt/stack/data/n-cpu-1/compute_id [ 537.056520] env[65121]: WARNING nova.compute.manager [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Compute nodes ['d56783bf-3ede-475a-8c5a-8d8303049e47'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 538.062762] env[65121]: INFO nova.compute.manager [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 539.068890] env[65121]: WARNING nova.compute.manager [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 539.069261] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 539.069300] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 539.069444] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 539.069588] env[65121]: DEBUG nova.compute.resource_tracker [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 539.070481] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a15ccbf-58f7-4ce2-806f-32c05fa2cc7d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.078874] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a43448b-9ba9-4588-9e54-b491c4aafc7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.094316] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6fd548-14fa-4fca-ad3d-1795a68ae873 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.101102] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f387b42a-eebf-4796-bb85-b2fd847238de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.129962] env[65121]: DEBUG nova.compute.resource_tracker [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180552MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 539.130262] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 539.130309] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 539.632600] env[65121]: WARNING nova.compute.resource_tracker [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] No compute node record for cpu-1:d56783bf-3ede-475a-8c5a-8d8303049e47: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host d56783bf-3ede-475a-8c5a-8d8303049e47 could not be found. [ 540.136137] env[65121]: INFO nova.compute.resource_tracker [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: d56783bf-3ede-475a-8c5a-8d8303049e47 [ 541.645226] env[65121]: DEBUG nova.compute.resource_tracker [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 541.645685] env[65121]: DEBUG nova.compute.resource_tracker [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 541.813487] env[65121]: INFO nova.scheduler.client.report [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] [req-e2043f15-0c18-4942-84a2-6851bd47a9ba] Created resource provider record via placement API for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 541.830975] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249afdae-8351-4bcf-a334-7da02fe94221 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.839218] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8166443-8f49-401b-89be-2a84e9e74641 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.869439] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bdf2c4-190a-464b-b35d-fb289178db62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.877593] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887c51b3-675b-4ebf-b8ed-128fdf7b1510 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.891527] env[65121]: DEBUG nova.compute.provider_tree [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 542.429944] env[65121]: DEBUG nova.scheduler.client.report [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 542.430250] env[65121]: DEBUG nova.compute.provider_tree [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 0 to 1 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 542.430397] env[65121]: DEBUG nova.compute.provider_tree [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 542.479949] env[65121]: DEBUG nova.compute.provider_tree [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 1 to 2 during operation: update_traits {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 542.985473] env[65121]: DEBUG nova.compute.resource_tracker [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 542.985862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.855s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 542.985862] env[65121]: DEBUG nova.service [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Creating RPC server for service compute {{(pid=65121) start /opt/stack/nova/nova/service.py:177}} [ 543.001051] env[65121]: DEBUG nova.service [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] Join ServiceGroup membership for this service compute {{(pid=65121) start /opt/stack/nova/nova/service.py:194}} [ 543.001237] env[65121]: DEBUG nova.servicegroup.drivers.db [None req-f06c0a2b-fe27-4483-9e8c-673dd684426d None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=65121) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 580.004224] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_power_states {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 580.508425] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Getting list of instances from cluster (obj){ [ 580.508425] env[65121]: value = "domain-c8" [ 580.508425] env[65121]: _type = "ClusterComputeResource" [ 580.508425] env[65121]: } {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 580.509878] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11204308-a057-4a35-8f52-5f3cf813335b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.520358] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Got total of 0 instances {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 580.520588] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 580.520902] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Getting list of instances from cluster (obj){ [ 580.520902] env[65121]: value = "domain-c8" [ 580.520902] env[65121]: _type = "ClusterComputeResource" [ 580.520902] env[65121]: } {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 580.522632] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e4936e-02d9-478e-83f1-fc338e781798 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.532999] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Got total of 0 instances {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 586.737943] env[65121]: INFO nova.utils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The default thread pool MainProcess.default is initialized [ 586.738863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 586.738863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 587.241392] env[65121]: DEBUG nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 587.792175] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 587.792569] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 587.795470] env[65121]: INFO nova.compute.claims [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.881444] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26916da-b1e3-474f-941b-eb340fd6c0d3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.896256] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c3ec97-bba3-44e6-982a-f47e7f0d49f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.930728] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cb28f3-398a-455d-ad44-f09f7d765ff9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.943562] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898e11da-c14a-46ef-a162-3101ea0728e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.959294] env[65121]: DEBUG nova.compute.provider_tree [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.984498] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "db737254-91fd-44bc-bf01-c864227cdf07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.985142] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "db737254-91fd-44bc-bf01-c864227cdf07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 589.465148] env[65121]: DEBUG nova.scheduler.client.report [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 589.489679] env[65121]: DEBUG nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 589.976021] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 589.976021] env[65121]: DEBUG nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 590.029298] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 590.031065] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 590.031390] env[65121]: INFO nova.compute.claims [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.048857] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "20396b81-2041-4bf2-85c4-8ea66c838ec4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 590.051108] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "20396b81-2041-4bf2-85c4-8ea66c838ec4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 590.486106] env[65121]: DEBUG nova.compute.utils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 590.488334] env[65121]: DEBUG nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 590.488591] env[65121]: DEBUG nova.network.neutron [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 590.489804] env[65121]: WARNING neutronclient.v2_0.client [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 590.491303] env[65121]: WARNING neutronclient.v2_0.client [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 590.492889] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 590.493658] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 590.558033] env[65121]: DEBUG nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 590.917116] env[65121]: DEBUG nova.policy [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f1ca61e513f4855a105e40ff37a2d75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '982f66705583488cb369f19160cc2ee5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 591.001108] env[65121]: DEBUG nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 591.090814] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 591.151663] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b657301-7f57-41d5-a148-7e99fb045971 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.164472] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2485696a-2479-4e25-a457-9edc43f7e0f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.202634] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a8f1ae2-9ed2-453e-b64a-6a71453198f3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.211147] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3287e607-8389-4e87-a56c-68257fc2ec33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.228264] env[65121]: DEBUG nova.compute.provider_tree [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.322715] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "0151ed85-bba4-443a-b18f-58272c973e1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 591.322936] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "0151ed85-bba4-443a-b18f-58272c973e1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 591.403692] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.404031] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.404262] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.404443] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.404621] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.404800] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.404975] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.405170] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 591.405318] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 591.437649] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "a43879c7-7378-4700-9581-8daabdafe5b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 591.438158] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "a43879c7-7378-4700-9581-8daabdafe5b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 591.578788] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 591.580042] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 591.718189] env[65121]: DEBUG nova.network.neutron [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Successfully created port: fc77c1b5-9d9b-479a-a553-e7961313af08 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 591.730399] env[65121]: DEBUG nova.scheduler.client.report [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 591.826949] env[65121]: DEBUG nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 591.908158] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 591.941435] env[65121]: DEBUG nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 592.013310] env[65121]: DEBUG nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 592.046592] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 592.046592] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 592.046592] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 592.046805] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 592.046954] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 592.047177] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 592.047364] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.047523] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 592.047927] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 592.048127] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 592.048311] env[65121]: DEBUG nova.virt.hardware [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 592.049398] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ca4c0b-e2d6-46f2-9381-59760c5f7623 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.058778] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fef3a1d-51a4-48ca-91cd-25a4d8b971a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.078141] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da2d1d2-0a50-435e-a286-915548694f33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.237022] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 592.237022] env[65121]: DEBUG nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 592.238942] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.148s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.240428] env[65121]: INFO nova.compute.claims [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.348838] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.464679] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.596810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.596810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.658612] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "aa9b6708-c53c-4117-9b75-9d506f393395" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.658870] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "aa9b6708-c53c-4117-9b75-9d506f393395" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.746233] env[65121]: DEBUG nova.compute.utils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 592.752710] env[65121]: DEBUG nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 592.752710] env[65121]: DEBUG nova.network.neutron [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 592.752710] env[65121]: WARNING neutronclient.v2_0.client [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 592.752710] env[65121]: WARNING neutronclient.v2_0.client [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 592.753228] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 592.754497] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 593.017407] env[65121]: DEBUG nova.policy [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd443765da2574f558da7dbdc0528f89f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a92a7a59dd5c44f9b925030ae81e344d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 593.099024] env[65121]: DEBUG nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 593.161418] env[65121]: DEBUG nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 593.262191] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 593.262191] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 593.265783] env[65121]: DEBUG nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 593.322155] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "5e915c8e-150a-4bfd-8005-a7a1a119057d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 593.322582] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 593.445258] env[65121]: DEBUG nova.network.neutron [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Successfully updated port: fc77c1b5-9d9b-479a-a553-e7961313af08 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 593.450352] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bf8061-3454-4b01-9344-01f53621a44c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.460965] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dd3289-9646-49f4-b42c-8719d82da413 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.498556] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ee204d-7915-42fa-800f-90b2acfc3df0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.508390] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9283da4b-8c1c-420a-b648-a9b267250c76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.524339] env[65121]: DEBUG nova.compute.provider_tree [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.623892] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 593.681545] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 593.765990] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 593.825666] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 593.929218] env[65121]: DEBUG nova.network.neutron [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Successfully created port: a059ff79-fc05-40c7-9def-3f3c476870b8 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 593.948692] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.948849] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 593.949043] env[65121]: DEBUG nova.network.neutron [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 594.027711] env[65121]: DEBUG nova.scheduler.client.report [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 594.278905] env[65121]: DEBUG nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 594.304784] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 594.320826] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 594.321036] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 594.321205] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 594.321379] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 594.321513] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 594.321648] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 594.321842] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.321986] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 594.322336] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 594.322497] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 594.323853] env[65121]: DEBUG nova.virt.hardware [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 594.323853] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1da225-5b80-49ea-8ac0-4fbd04845a8d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.336752] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0823a530-d50d-495b-927e-a1d9ae8a0798 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.356152] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 594.453907] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 594.454998] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.533705] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 594.534347] env[65121]: DEBUG nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 594.537393] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.629s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 594.537393] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 594.537511] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 594.537747] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.189s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 594.539134] env[65121]: INFO nova.compute.claims [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.546423] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4867c3dd-2f69-4310-b026-66505aed1411 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.560092] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701563b9-f207-45c6-849d-c24ed0be6ddf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.579413] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e832c2ce-e310-443c-a243-02adbfbd908c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.587439] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cb7b45-74bf-4cb3-ad44-80c56ea89237 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.623586] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180534MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 594.623847] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 594.822589] env[65121]: DEBUG nova.network.neutron [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 595.048653] env[65121]: DEBUG nova.compute.utils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 595.050099] env[65121]: DEBUG nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 595.447174] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 595.447174] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 595.514350] env[65121]: DEBUG nova.network.neutron [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Successfully updated port: a059ff79-fc05-40c7-9def-3f3c476870b8 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 595.552197] env[65121]: DEBUG nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 595.749335] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371d04c8-8477-4a3c-89f1-6ae022f7e292 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.758220] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a96da43-3067-4c15-8114-e14520a722f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.795293] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4628ba4d-b951-4630-8342-442756918cb8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.804983] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142beaab-8cc1-466f-9378-fe154b6ef6a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.820552] env[65121]: DEBUG nova.compute.provider_tree [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.856850] env[65121]: DEBUG nova.compute.manager [req-b6ca32d2-1a04-4627-bb97-7ad77908ca92 req-3fb381af-f155-454a-9300-88a9b0464fd0 service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Received event network-vif-plugged-fc77c1b5-9d9b-479a-a553-e7961313af08 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 595.858078] env[65121]: DEBUG oslo_concurrency.lockutils [req-b6ca32d2-1a04-4627-bb97-7ad77908ca92 req-3fb381af-f155-454a-9300-88a9b0464fd0 service nova] Acquiring lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 595.858078] env[65121]: DEBUG oslo_concurrency.lockutils [req-b6ca32d2-1a04-4627-bb97-7ad77908ca92 req-3fb381af-f155-454a-9300-88a9b0464fd0 service nova] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 595.858078] env[65121]: DEBUG oslo_concurrency.lockutils [req-b6ca32d2-1a04-4627-bb97-7ad77908ca92 req-3fb381af-f155-454a-9300-88a9b0464fd0 service nova] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 595.858359] env[65121]: DEBUG nova.compute.manager [req-b6ca32d2-1a04-4627-bb97-7ad77908ca92 req-3fb381af-f155-454a-9300-88a9b0464fd0 service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] No waiting events found dispatching network-vif-plugged-fc77c1b5-9d9b-479a-a553-e7961313af08 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 595.858431] env[65121]: WARNING nova.compute.manager [req-b6ca32d2-1a04-4627-bb97-7ad77908ca92 req-3fb381af-f155-454a-9300-88a9b0464fd0 service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Received unexpected event network-vif-plugged-fc77c1b5-9d9b-479a-a553-e7961313af08 for instance with vm_state building and task_state spawning. [ 595.914603] env[65121]: WARNING neutronclient.v2_0.client [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 595.915260] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 595.915585] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.017929] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "refresh_cache-db737254-91fd-44bc-bf01-c864227cdf07" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.018223] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired lock "refresh_cache-db737254-91fd-44bc-bf01-c864227cdf07" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 596.018438] env[65121]: DEBUG nova.network.neutron [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 596.326433] env[65121]: DEBUG nova.scheduler.client.report [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 596.409592] env[65121]: DEBUG nova.network.neutron [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance_info_cache with network_info: [{"id": "fc77c1b5-9d9b-479a-a553-e7961313af08", "address": "fa:16:3e:f6:e5:10", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.52", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc77c1b5-9d", "ovs_interfaceid": "fc77c1b5-9d9b-479a-a553-e7961313af08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 596.523429] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.523819] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.566934] env[65121]: DEBUG nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 596.612816] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 596.613339] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 596.613590] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 596.613857] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 596.614084] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 596.614289] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 596.614574] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.614796] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 596.615332] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 596.615332] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 596.619366] env[65121]: DEBUG nova.virt.hardware [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 596.620670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6dae7a6-bad3-4ea5-8577-300af6c6a83d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.634331] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f50d689-6c4e-4f98-8677-377738078d8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.654055] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 596.662973] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 596.664260] env[65121]: DEBUG nova.network.neutron [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 596.667039] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe9c60a0-1913-41e2-9952-a88aea44eb9e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.682159] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Created folder: OpenStack in parent group-v4. [ 596.682384] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Creating folder: Project (21027ead4f1042029f75e43c3e015eb8). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 596.682874] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08c59a25-1e90-4459-a12f-8e85cdddff8b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.694812] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Created folder: Project (21027ead4f1042029f75e43c3e015eb8) in parent group-v993268. [ 596.694812] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Creating folder: Instances. Parent ref: group-v993269. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 596.694812] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04827935-6ca5-4105-9b09-e79bef60082c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.707971] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Created folder: Instances in parent group-v993269. [ 596.707971] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 596.708087] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 596.708287] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b94d9d9-d3a2-4f16-84d8-df81e2e43cef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.727333] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 596.727333] env[65121]: value = "task-5105873" [ 596.727333] env[65121]: _type = "Task" [ 596.727333] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.729260] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.729260] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.748980] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105873, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.833497] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 596.834603] env[65121]: DEBUG nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 596.839227] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.375s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 596.842554] env[65121]: INFO nova.compute.claims [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.890019] env[65121]: WARNING neutronclient.v2_0.client [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 596.890390] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.890515] env[65121]: WARNING openstack [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.913658] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 596.915418] env[65121]: DEBUG nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Instance network_info: |[{"id": "fc77c1b5-9d9b-479a-a553-e7961313af08", "address": "fa:16:3e:f6:e5:10", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.52", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc77c1b5-9d", "ovs_interfaceid": "fc77c1b5-9d9b-479a-a553-e7961313af08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 596.917594] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:e5:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fc77c1b5-9d9b-479a-a553-e7961313af08', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 596.932916] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Creating folder: Project (982f66705583488cb369f19160cc2ee5). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 596.938909] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35d6322d-3635-4da0-9037-17adacff7cdc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.950020] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Created folder: Project (982f66705583488cb369f19160cc2ee5) in parent group-v993268. [ 596.950240] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Creating folder: Instances. Parent ref: group-v993272. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 596.950604] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-365b80b2-2183-45f5-89e9-683f4485b050 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.963052] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Created folder: Instances in parent group-v993272. [ 596.963052] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 596.963052] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 596.963052] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e7ef37f-db0f-4e31-9605-4fe2d84336ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.998675] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 596.998675] env[65121]: value = "task-5105876" [ 596.998675] env[65121]: _type = "Task" [ 596.998675] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.012152] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105876, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.075204] env[65121]: DEBUG nova.network.neutron [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Updating instance_info_cache with network_info: [{"id": "a059ff79-fc05-40c7-9def-3f3c476870b8", "address": "fa:16:3e:0a:65:44", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.124", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa059ff79-fc", "ovs_interfaceid": "a059ff79-fc05-40c7-9def-3f3c476870b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 597.262640] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105873, 'name': CreateVM_Task, 'duration_secs': 0.318766} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.262833] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 597.263960] env[65121]: DEBUG oslo_vmware.service [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c0c09b-51b3-4442-82a6-98dd3b96d80b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.273324] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.273324] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 597.275603] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 597.275603] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0b30808-59c3-40c5-998a-c43a26dd8de8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.286288] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 597.286288] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522942a9-32ea-f53e-c2d1-397aa220d97b" [ 597.286288] env[65121]: _type = "Task" [ 597.286288] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.312652] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 597.313129] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 597.314934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.315055] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 597.316727] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 597.317103] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5469b27-585b-40af-9575-2ab9292e23f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.340360] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 597.340738] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 597.346660] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2224c322-d602-4576-bb5f-536b81dad98d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.358057] env[65121]: DEBUG nova.compute.utils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 597.378416] env[65121]: DEBUG nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 597.387754] env[65121]: DEBUG nova.network.neutron [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 597.387754] env[65121]: WARNING neutronclient.v2_0.client [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.387754] env[65121]: WARNING neutronclient.v2_0.client [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.388147] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 597.388754] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 597.413947] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0ad3a10-88aa-4be6-ad83-4a1ea94b8071 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.435124] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 597.435124] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52342d37-defe-43f2-6682-4dbb34356806" [ 597.435124] env[65121]: _type = "Task" [ 597.435124] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.459444] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Preparing fetch location {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 597.459444] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Creating directory with path [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 597.459605] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21a199c8-0970-4b26-889f-5cb9593d7c15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.482423] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Created directory with path [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 597.482609] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Fetch image to [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 597.482669] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Downloading image file data 3ed5d656-a616-4609-8bb7-b02beb3ac3df to [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk on the data store datastore1 {{(pid=65121) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 597.483910] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c14ccaa-d79f-4e96-b31d-5a8a45468154 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.498801] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525e712d-4571-4d11-a4f5-fff9d478a47b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.516105] env[65121]: DEBUG nova.policy [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a09ed33e751e4039a05bdfb907fc7ce2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d44b5f8148445c8bbaa5fd09032c0c0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 597.529722] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755c3f6a-276b-4a2f-996b-3b2f4554a5e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.577037] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105876, 'name': CreateVM_Task, 'duration_secs': 0.42953} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.578187] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Releasing lock "refresh_cache-db737254-91fd-44bc-bf01-c864227cdf07" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 597.578187] env[65121]: DEBUG nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Instance network_info: |[{"id": "a059ff79-fc05-40c7-9def-3f3c476870b8", "address": "fa:16:3e:0a:65:44", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.124", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa059ff79-fc", "ovs_interfaceid": "a059ff79-fc05-40c7-9def-3f3c476870b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 597.578536] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 597.579282] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:65:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a059ff79-fc05-40c7-9def-3f3c476870b8', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 597.586172] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Creating folder: Project (a92a7a59dd5c44f9b925030ae81e344d). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 597.587081] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c4fcc2-ae85-4e74-a4b0-f8fc9fb1fdc6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.590991] env[65121]: WARNING neutronclient.v2_0.client [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 597.590991] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 597.590991] env[65121]: WARNING openstack [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 597.599341] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4f1f6b2-90f8-420d-a9e4-dec3f5d9220d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.606533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 597.606658] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "cc002961-d742-4255-88c7-f0a5a39424b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 597.609927] env[65121]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-07b8b32f-f102-4dcc-868f-1cddd9c176df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.612311] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Created folder: Project (a92a7a59dd5c44f9b925030ae81e344d) in parent group-v993268. [ 597.612696] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Creating folder: Instances. Parent ref: group-v993275. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 597.612696] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0878be5e-f312-4a57-a73d-f647fdb5053d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.624243] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Created folder: Instances in parent group-v993275. [ 597.624243] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 597.624369] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 597.624631] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb80f2f1-7002-4c95-89fb-6a06a5aad1ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.650678] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Downloading image file data 3ed5d656-a616-4609-8bb7-b02beb3ac3df to the data store datastore1 {{(pid=65121) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 597.657059] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 597.657059] env[65121]: value = "task-5105879" [ 597.657059] env[65121]: _type = "Task" [ 597.657059] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.670074] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105879, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.819707] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.819707] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 597.820092] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 597.821290] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e51052bf-44f8-48d7-ab9c-132f073135bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.824626] env[65121]: DEBUG oslo_vmware.rw_handles [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65121) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 597.879623] env[65121]: DEBUG nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 597.889226] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 597.889226] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52eb7dc2-b6af-7262-a7e3-5e3ff1f175d6" [ 597.889226] env[65121]: _type = "Task" [ 597.889226] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.902734] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 597.902734] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 597.902734] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.099682] env[65121]: DEBUG nova.network.neutron [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Successfully created port: cf0d4936-618c-45a3-af9c-3f745b2cf316 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 598.109932] env[65121]: DEBUG nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 598.147922] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9d7575-d093-4614-87d5-23bbc9840116 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.162712] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a071b1d2-dbad-4037-9b56-8260edb0bef8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.175776] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105879, 'name': CreateVM_Task, 'duration_secs': 0.371184} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.200822] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 598.205058] env[65121]: WARNING neutronclient.v2_0.client [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 598.205230] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.206122] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 598.206122] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 598.206527] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0657f6d6-dce6-4fef-8352-b244e9a50cf3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.209331] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddc438b2-339f-4c86-936f-ac6f03d00a86 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.217732] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 598.217732] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527aa714-ddc5-a996-7a6e-e12f78fbef70" [ 598.217732] env[65121]: _type = "Task" [ 598.217732] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.226383] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61df5151-3c4b-45fc-a1d9-d3cf852c5ebf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.238797] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527aa714-ddc5-a996-7a6e-e12f78fbef70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.249260] env[65121]: DEBUG nova.compute.provider_tree [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.258487] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "f1a5fdce-a940-43d4-83d7-716786c9ac34" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 598.258652] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 598.641226] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 598.734966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 598.735261] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 598.735516] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.758801] env[65121]: DEBUG nova.scheduler.client.report [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 598.764507] env[65121]: DEBUG oslo_vmware.rw_handles [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Completed reading data from the image iterator. {{(pid=65121) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 598.764725] env[65121]: DEBUG oslo_vmware.rw_handles [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 598.842735] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Downloaded image file data 3ed5d656-a616-4609-8bb7-b02beb3ac3df to vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk on the data store datastore1 {{(pid=65121) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 598.844306] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Caching image {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 598.844550] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Copying Virtual Disk [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk to [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 598.845749] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97f02568-d5cc-4800-b5c6-c1abca0d3c23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.853950] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 598.853950] env[65121]: value = "task-5105880" [ 598.853950] env[65121]: _type = "Task" [ 598.853950] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.862363] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105880, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.897269] env[65121]: DEBUG nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 598.933675] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 598.934086] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 598.934335] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 598.934602] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 598.934834] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 598.935078] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 598.935390] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.935595] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 598.935810] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 598.936058] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 598.936297] env[65121]: DEBUG nova.virt.hardware [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 598.941276] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1c6c2b-ffc3-47e6-a2f2-89f67d80c1f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.952518] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623933ff-1474-436f-8597-d5c61cde4ffe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.264577] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 599.265131] env[65121]: DEBUG nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 599.267973] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.644s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.269439] env[65121]: INFO nova.compute.claims [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.369905] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105880, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.777770] env[65121]: DEBUG nova.compute.utils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 599.780869] env[65121]: DEBUG nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 599.781330] env[65121]: DEBUG nova.network.neutron [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 599.781424] env[65121]: WARNING neutronclient.v2_0.client [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.781702] env[65121]: WARNING neutronclient.v2_0.client [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.782292] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.782611] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.864778] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105880, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690017} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.865034] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Copied Virtual Disk [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk to [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 599.865208] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleting the datastore file [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 599.865451] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1437b4b0-c593-4add-af70-df68e32878e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.872344] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 599.872344] env[65121]: value = "task-5105881" [ 599.872344] env[65121]: _type = "Task" [ 599.872344] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.882529] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105881, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.034685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "6887bb66-7186-4d58-be59-d0c53c3221c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.034685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "6887bb66-7186-4d58-be59-d0c53c3221c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 600.281925] env[65121]: DEBUG nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 600.381981] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105881, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029263} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.384345] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 600.384345] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Moving file from [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046/3ed5d656-a616-4609-8bb7-b02beb3ac3df to [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df. {{(pid=65121) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 600.384345] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-f849a792-fe0f-40c0-9461-557355d22a23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.393381] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 600.393381] env[65121]: value = "task-5105882" [ 600.393381] env[65121]: _type = "Task" [ 600.393381] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.408113] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105882, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.464671] env[65121]: DEBUG nova.network.neutron [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Successfully updated port: cf0d4936-618c-45a3-af9c-3f745b2cf316 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 600.511505] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd273b3-8cee-41d2-8a45-41c711effc1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.522044] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc6413c1-7747-4991-85c5-a5cb77a73b4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.569975] env[65121]: DEBUG nova.policy [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '390615f3a68d489183e81f8c888e5247', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '018269ed0c644392aebdb56cf0523292', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 600.572203] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7249604-31fd-4996-9829-8b20c69054f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.581237] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e8724e-4c2d-4f94-92fd-2078bc832b98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.595522] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 600.906781] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105882, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.027523} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.907418] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] File moved {{(pid=65121) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 600.907418] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Cleaning up location [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 600.907418] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleting the datastore file [datastore1] vmware_temp/6dd3b2c1-c6bc-4570-bb4f-b92de9865046 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 600.907545] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb539ba1-4296-4284-8343-f1ebdd4bf657 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.915996] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 600.915996] env[65121]: value = "task-5105883" [ 600.915996] env[65121]: _type = "Task" [ 600.915996] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.928749] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105883, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.967730] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "refresh_cache-0151ed85-bba4-443a-b18f-58272c973e1e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.969074] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquired lock "refresh_cache-0151ed85-bba4-443a-b18f-58272c973e1e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 600.969074] env[65121]: DEBUG nova.network.neutron [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 601.034272] env[65121]: DEBUG nova.network.neutron [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Successfully created port: e6d39f1f-d115-4218-a28a-e2cd0d05868d {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 601.138325] env[65121]: ERROR nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [req-54cdf1bf-8bb2-4d05-b9ca-943c34fe8fd7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-54cdf1bf-8bb2-4d05-b9ca-943c34fe8fd7"}]} [ 601.161791] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 601.182568] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 601.183147] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 601.205849] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: 01907032-0366-4fa9-a913-69845a5eac36 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 601.253420] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 601.253657] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 15 to 16 during operation: update_traits {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 601.303142] env[65121]: DEBUG nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 601.341618] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 601.341837] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 601.341924] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 601.342165] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 601.343687] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 601.343687] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 601.343687] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.343687] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 601.343687] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 601.343890] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 601.343890] env[65121]: DEBUG nova.virt.hardware [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 601.344356] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69fa493-7109-411b-9337-f12c03aa0666 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.356499] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60da3949-f217-437d-b3ca-b0b948a72796 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.429554] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105883, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027355} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.429793] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 601.430577] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-604f1d22-cbc0-4e64-9018-b0118657f510 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.444627] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 601.444627] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522c4c80-2ee3-49aa-9c48-318264eee26d" [ 601.444627] env[65121]: _type = "Task" [ 601.444627] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.458121] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522c4c80-2ee3-49aa-9c48-318264eee26d, 'name': SearchDatastore_Task, 'duration_secs': 0.00908} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.458498] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.458733] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 601.459078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.459277] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 601.459486] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9cde42a1-4d2b-474f-b258-be53fff4f700 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.462317] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddb39cb8-6f59-4f3d-a82e-a9eef5041f44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.471632] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.471970] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.487188] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 601.487188] env[65121]: value = "task-5105884" [ 601.487188] env[65121]: _type = "Task" [ 601.487188] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.489152] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 601.489152] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 601.489582] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fd60e2c-66eb-42cf-a93c-2a1cac945c99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.502499] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 601.502499] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c3fed6-f2e3-8541-998c-6ac91a52bb67" [ 601.502499] env[65121]: _type = "Task" [ 601.502499] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.505643] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.515778] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c3fed6-f2e3-8541-998c-6ac91a52bb67, 'name': SearchDatastore_Task, 'duration_secs': 0.008581} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.517124] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-018e5e7d-47ab-4e82-9686-2fd4d9d7606b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.525277] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 601.525277] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52331014-8101-21b5-9df5-cd3a3a45fa31" [ 601.525277] env[65121]: _type = "Task" [ 601.525277] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.537941] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52331014-8101-21b5-9df5-cd3a3a45fa31, 'name': SearchDatastore_Task, 'duration_secs': 0.009391} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.538260] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.538692] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 5f6e4f46-0745-42c0-a779-6cffb60e21ca/5f6e4f46-0745-42c0-a779-6cffb60e21ca.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 601.538992] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.539202] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 601.539412] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88a6ea38-df98-4ab0-a2eb-7917207e0444 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.544631] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4d109b4-7f33-4504-8112-4601c4cbbc8b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.554953] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 601.554953] env[65121]: value = "task-5105885" [ 601.554953] env[65121]: _type = "Task" [ 601.554953] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.555963] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 601.555963] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 601.560310] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-985ea3d8-d2d0-4db3-a865-7c85ddc229a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.569939] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105885, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.574017] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 601.574017] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ce2c8a-5302-7040-0e2f-a2ae9f7ee324" [ 601.574017] env[65121]: _type = "Task" [ 601.574017] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.583438] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ce2c8a-5302-7040-0e2f-a2ae9f7ee324, 'name': SearchDatastore_Task, 'duration_secs': 0.009166} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.584451] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-661da533-6145-451c-9efe-fe341da567bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.591644] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 601.591644] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52423eeb-4f5c-c25c-3aa4-5876e9a86468" [ 601.591644] env[65121]: _type = "Task" [ 601.591644] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.595356] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7cd351c-0218-452a-9744-b390a80a4336 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.606899] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49ed0ae-551d-440f-9a39-2cba8d2c1baa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.612114] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52423eeb-4f5c-c25c-3aa4-5876e9a86468, 'name': SearchDatastore_Task, 'duration_secs': 0.010002} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.613024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.613024] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] db737254-91fd-44bc-bf01-c864227cdf07/db737254-91fd-44bc-bf01-c864227cdf07.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 601.613355] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2c57c0a-73e2-4f0a-802d-b01d6321b20e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.646255] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc034ce-c722-42d6-acdb-273ab2e29420 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.649040] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 601.649040] env[65121]: value = "task-5105886" [ 601.649040] env[65121]: _type = "Task" [ 601.649040] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.659550] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83ef83b-a558-4bf1-8d95-00b7909b0d47 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.664537] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.675754] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 601.685169] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 601.686355] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 601.686355] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 601.686483] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 601.695440] env[65121]: DEBUG nova.network.neutron [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 601.940317] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.940876] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.003552] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105884, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.027072] env[65121]: DEBUG nova.compute.manager [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Received event network-changed-fc77c1b5-9d9b-479a-a553-e7961313af08 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 602.027252] env[65121]: DEBUG nova.compute.manager [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Refreshing instance network info cache due to event network-changed-fc77c1b5-9d9b-479a-a553-e7961313af08. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 602.027457] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Acquiring lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.027597] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Acquired lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 602.027726] env[65121]: DEBUG nova.network.neutron [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Refreshing network info cache for port fc77c1b5-9d9b-479a-a553-e7961313af08 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 602.070298] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105885, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.160838] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.207367] env[65121]: ERROR nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [req-7c82b07e-7903-4e05-a47d-dd4cab8f9830] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7c82b07e-7903-4e05-a47d-dd4cab8f9830"}]} [ 602.228262] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 602.245837] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 602.246089] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 602.259999] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 602.286067] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 602.508816] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532965} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.509166] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 602.509382] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 602.509668] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2976150-8a00-490c-a315-4029e7e6242f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.520818] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 602.520818] env[65121]: value = "task-5105887" [ 602.520818] env[65121]: _type = "Task" [ 602.520818] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.525650] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2a9237-2893-4bcc-b73e-9683a605fe8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.531335] env[65121]: WARNING neutronclient.v2_0.client [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.532025] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.532292] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.546539] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105887, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.547991] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c7d456-709f-4a8f-b8de-004b717bccba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.589936] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e269f9a-c56b-4e3a-8aa2-395d2eda9bdd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.603461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d963afb4-742f-4689-aed8-c4f716297610 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.608793] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105885, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.838934} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.609079] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 5f6e4f46-0745-42c0-a779-6cffb60e21ca/5f6e4f46-0745-42c0-a779-6cffb60e21ca.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 602.609291] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 602.609958] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23624810-492d-4bd0-9e80-539004e5fa85 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.622455] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 602.625839] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 602.625839] env[65121]: value = "task-5105888" [ 602.625839] env[65121]: _type = "Task" [ 602.625839] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.639308] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105888, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.665743] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105886, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.686271] env[65121]: DEBUG nova.network.neutron [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Successfully updated port: e6d39f1f-d115-4218-a28a-e2cd0d05868d {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 602.851302] env[65121]: WARNING neutronclient.v2_0.client [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.852097] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.852727] env[65121]: WARNING openstack [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.034986] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105887, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103222} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.035332] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 603.036645] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53341f13-f2d2-4ea7-8ecd-6e10d1beafe0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.058412] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 603.058721] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-348153f1-2dea-4c8c-a788-d601f98f9037 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.083528] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 603.083528] env[65121]: value = "task-5105889" [ 603.083528] env[65121]: _type = "Task" [ 603.083528] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.094394] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105889, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.139372] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105888, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089508} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.139372] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 603.140488] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18ba6b4-453b-40d2-9049-5fdb19164d3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.173057] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 5f6e4f46-0745-42c0-a779-6cffb60e21ca/5f6e4f46-0745-42c0-a779-6cffb60e21ca.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 603.177377] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47ddf26b-c3cd-496c-a59e-8c73ef456cde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.196138] env[65121]: DEBUG nova.scheduler.client.report [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 18 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 603.196282] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 18 to 19 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 603.196315] env[65121]: DEBUG nova.compute.provider_tree [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 603.200479] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "refresh_cache-a43879c7-7378-4700-9581-8daabdafe5b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.201202] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired lock "refresh_cache-a43879c7-7378-4700-9581-8daabdafe5b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 603.201202] env[65121]: DEBUG nova.network.neutron [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 603.208774] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105886, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.236129} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.210319] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] db737254-91fd-44bc-bf01-c864227cdf07/db737254-91fd-44bc-bf01-c864227cdf07.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 603.210576] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 603.210917] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 603.210917] env[65121]: value = "task-5105890" [ 603.210917] env[65121]: _type = "Task" [ 603.210917] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.211786] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0af2ebe-f6a5-403d-ba34-2b96f16096d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.225166] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105890, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.225572] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 603.225572] env[65121]: value = "task-5105891" [ 603.225572] env[65121]: _type = "Task" [ 603.225572] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.237191] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105891, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.274925] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.275354] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.378311] env[65121]: DEBUG nova.network.neutron [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Updating instance_info_cache with network_info: [{"id": "cf0d4936-618c-45a3-af9c-3f745b2cf316", "address": "fa:16:3e:88:13:d9", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf0d4936-61", "ovs_interfaceid": "cf0d4936-618c-45a3-af9c-3f745b2cf316", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 603.569824] env[65121]: WARNING neutronclient.v2_0.client [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.570531] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.571041] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.594651] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105889, 'name': ReconfigVM_Task, 'duration_secs': 0.336332} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.595162] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 603.596021] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1311895-57b9-4b05-ac10-52b3fd4ecb02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.605799] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 603.605799] env[65121]: value = "task-5105892" [ 603.605799] env[65121]: _type = "Task" [ 603.605799] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.621176] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105892, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.703799] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.436s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 603.704349] env[65121]: DEBUG nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 603.708277] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.708404] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.725464] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.044s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 603.727784] env[65121]: INFO nova.compute.claims [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.747231] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105890, 'name': ReconfigVM_Task, 'duration_secs': 0.327816} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.750463] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 5f6e4f46-0745-42c0-a779-6cffb60e21ca/5f6e4f46-0745-42c0-a779-6cffb60e21ca.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 603.751264] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105891, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079813} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.751471] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f60699a7-38c8-45b6-b185-a1fb76811b54 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.753449] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 603.754539] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c279106-e6bf-4aa3-b5a8-e65c00fb07a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.761783] env[65121]: DEBUG nova.network.neutron [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 603.777931] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 603.777931] env[65121]: value = "task-5105893" [ 603.777931] env[65121]: _type = "Task" [ 603.777931] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.788051] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] db737254-91fd-44bc-bf01-c864227cdf07/db737254-91fd-44bc-bf01-c864227cdf07.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 603.789591] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.790190] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.801563] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-330a8898-d3b0-4c63-a2b2-16720ba45c66 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.818842] env[65121]: DEBUG nova.network.neutron [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updated VIF entry in instance network info cache for port fc77c1b5-9d9b-479a-a553-e7961313af08. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 603.819206] env[65121]: DEBUG nova.network.neutron [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance_info_cache with network_info: [{"id": "fc77c1b5-9d9b-479a-a553-e7961313af08", "address": "fa:16:3e:f6:e5:10", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.52", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc77c1b5-9d", "ovs_interfaceid": "fc77c1b5-9d9b-479a-a553-e7961313af08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 603.831760] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105893, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.833656] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 603.833656] env[65121]: value = "task-5105894" [ 603.833656] env[65121]: _type = "Task" [ 603.833656] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.847042] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105894, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.879543] env[65121]: WARNING neutronclient.v2_0.client [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.880311] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.880568] env[65121]: WARNING openstack [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.889624] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Releasing lock "refresh_cache-0151ed85-bba4-443a-b18f-58272c973e1e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 603.889833] env[65121]: DEBUG nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Instance network_info: |[{"id": "cf0d4936-618c-45a3-af9c-3f745b2cf316", "address": "fa:16:3e:88:13:d9", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf0d4936-61", "ovs_interfaceid": "cf0d4936-618c-45a3-af9c-3f745b2cf316", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 603.892726] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:13:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf0d4936-618c-45a3-af9c-3f745b2cf316', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 603.900480] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Creating folder: Project (1d44b5f8148445c8bbaa5fd09032c0c0). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 603.902701] env[65121]: DEBUG nova.compute.manager [req-3b1cd786-6357-4d29-a1dc-5e65a4672c98 req-cfefda46-69ac-4d21-adf9-2124e33bf927 service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Received event network-vif-plugged-cf0d4936-618c-45a3-af9c-3f745b2cf316 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 603.902701] env[65121]: DEBUG oslo_concurrency.lockutils [req-3b1cd786-6357-4d29-a1dc-5e65a4672c98 req-cfefda46-69ac-4d21-adf9-2124e33bf927 service nova] Acquiring lock "0151ed85-bba4-443a-b18f-58272c973e1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 603.902701] env[65121]: DEBUG oslo_concurrency.lockutils [req-3b1cd786-6357-4d29-a1dc-5e65a4672c98 req-cfefda46-69ac-4d21-adf9-2124e33bf927 service nova] Lock "0151ed85-bba4-443a-b18f-58272c973e1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 603.902701] env[65121]: DEBUG oslo_concurrency.lockutils [req-3b1cd786-6357-4d29-a1dc-5e65a4672c98 req-cfefda46-69ac-4d21-adf9-2124e33bf927 service nova] Lock "0151ed85-bba4-443a-b18f-58272c973e1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 603.902701] env[65121]: DEBUG nova.compute.manager [req-3b1cd786-6357-4d29-a1dc-5e65a4672c98 req-cfefda46-69ac-4d21-adf9-2124e33bf927 service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] No waiting events found dispatching network-vif-plugged-cf0d4936-618c-45a3-af9c-3f745b2cf316 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 603.903262] env[65121]: WARNING nova.compute.manager [req-3b1cd786-6357-4d29-a1dc-5e65a4672c98 req-cfefda46-69ac-4d21-adf9-2124e33bf927 service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Received unexpected event network-vif-plugged-cf0d4936-618c-45a3-af9c-3f745b2cf316 for instance with vm_state building and task_state spawning. [ 603.903408] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84771d10-e391-497b-8326-3232503b4822 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.918161] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Created folder: Project (1d44b5f8148445c8bbaa5fd09032c0c0) in parent group-v993268. [ 603.918161] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Creating folder: Instances. Parent ref: group-v993278. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 603.918161] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fcb6c656-45f3-42af-9ed0-afb5bf84d3b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.929434] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Created folder: Instances in parent group-v993278. [ 603.929833] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 603.930044] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 603.930301] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6103596d-5b59-4a30-8a04-cd1b94d4eba2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.953985] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 603.953985] env[65121]: value = "task-5105897" [ 603.953985] env[65121]: _type = "Task" [ 603.953985] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.970682] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105897, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.008125] env[65121]: DEBUG nova.network.neutron [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Updating instance_info_cache with network_info: [{"id": "e6d39f1f-d115-4218-a28a-e2cd0d05868d", "address": "fa:16:3e:4b:6e:b3", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.158", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6d39f1f-d1", "ovs_interfaceid": "e6d39f1f-d115-4218-a28a-e2cd0d05868d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 604.117819] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105892, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.220812] env[65121]: DEBUG nova.compute.utils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 604.221760] env[65121]: DEBUG nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 604.221946] env[65121]: DEBUG nova.network.neutron [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 604.225199] env[65121]: WARNING neutronclient.v2_0.client [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.225606] env[65121]: WARNING neutronclient.v2_0.client [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.226189] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 604.226613] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 604.303561] env[65121]: DEBUG nova.policy [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e773a3343a448f5acd3f08005e8ed45', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '40aee06073b04c25bc7b38fbd09856e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 604.317646] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105893, 'name': Rename_Task, 'duration_secs': 0.325184} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.318726] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 604.318726] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3df995a5-9c9d-4a37-b832-758bbaa6c7f6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.326430] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Releasing lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 604.326430] env[65121]: DEBUG nova.compute.manager [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Received event network-vif-plugged-a059ff79-fc05-40c7-9def-3f3c476870b8 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 604.326430] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Acquiring lock "db737254-91fd-44bc-bf01-c864227cdf07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 604.326706] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Lock "db737254-91fd-44bc-bf01-c864227cdf07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 604.326706] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Lock "db737254-91fd-44bc-bf01-c864227cdf07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 604.326890] env[65121]: DEBUG nova.compute.manager [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] No waiting events found dispatching network-vif-plugged-a059ff79-fc05-40c7-9def-3f3c476870b8 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 604.326970] env[65121]: WARNING nova.compute.manager [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Received unexpected event network-vif-plugged-a059ff79-fc05-40c7-9def-3f3c476870b8 for instance with vm_state building and task_state spawning. [ 604.327286] env[65121]: DEBUG nova.compute.manager [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Received event network-changed-a059ff79-fc05-40c7-9def-3f3c476870b8 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 604.327436] env[65121]: DEBUG nova.compute.manager [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Refreshing instance network info cache due to event network-changed-a059ff79-fc05-40c7-9def-3f3c476870b8. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 604.327605] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Acquiring lock "refresh_cache-db737254-91fd-44bc-bf01-c864227cdf07" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.327723] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Acquired lock "refresh_cache-db737254-91fd-44bc-bf01-c864227cdf07" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 604.327890] env[65121]: DEBUG nova.network.neutron [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Refreshing network info cache for port a059ff79-fc05-40c7-9def-3f3c476870b8 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 604.329197] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 604.329197] env[65121]: value = "task-5105898" [ 604.329197] env[65121]: _type = "Task" [ 604.329197] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.341259] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105898, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.348673] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105894, 'name': ReconfigVM_Task, 'duration_secs': 0.500742} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.348673] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Reconfigured VM instance instance-00000002 to attach disk [datastore1] db737254-91fd-44bc-bf01-c864227cdf07/db737254-91fd-44bc-bf01-c864227cdf07.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 604.352038] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2b7a01f-d032-40ab-b384-57ae1de0c6fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.357033] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 604.357033] env[65121]: value = "task-5105899" [ 604.357033] env[65121]: _type = "Task" [ 604.357033] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.370528] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105899, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.473890] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105897, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.511569] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Releasing lock "refresh_cache-a43879c7-7378-4700-9581-8daabdafe5b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 604.511726] env[65121]: DEBUG nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Instance network_info: |[{"id": "e6d39f1f-d115-4218-a28a-e2cd0d05868d", "address": "fa:16:3e:4b:6e:b3", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.158", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6d39f1f-d1", "ovs_interfaceid": "e6d39f1f-d115-4218-a28a-e2cd0d05868d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 604.512243] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:6e:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e6d39f1f-d115-4218-a28a-e2cd0d05868d', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 604.520187] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Creating folder: Project (018269ed0c644392aebdb56cf0523292). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.520551] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f782296-64ca-4561-8e5c-5ba5a03d9152 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.533884] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Created folder: Project (018269ed0c644392aebdb56cf0523292) in parent group-v993268. [ 604.534355] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Creating folder: Instances. Parent ref: group-v993281. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 604.537497] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79762054-bc5b-46e3-ab59-ed1ebc58b0e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.544340] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Created folder: Instances in parent group-v993281. [ 604.544631] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 604.544766] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 604.544981] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51120f06-a46e-4d01-8140-fc63dffc1641 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.568192] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 604.568192] env[65121]: value = "task-5105902" [ 604.568192] env[65121]: _type = "Task" [ 604.568192] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.575994] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105902, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.616433] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105892, 'name': Rename_Task, 'duration_secs': 0.924521} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.616711] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 604.617047] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8170cde9-f04d-4a65-aefd-394b4e2cd490 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.624300] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 604.624300] env[65121]: value = "task-5105903" [ 604.624300] env[65121]: _type = "Task" [ 604.624300] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.633459] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105903, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.702889] env[65121]: DEBUG nova.network.neutron [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Successfully created port: 7aef5809-0a17-4d3e-b553-93f337e59e89 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 604.735434] env[65121]: DEBUG nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 604.837228] env[65121]: WARNING neutronclient.v2_0.client [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.838267] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 604.838808] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 604.873026] env[65121]: DEBUG oslo_vmware.api [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5105898, 'name': PowerOnVM_Task, 'duration_secs': 0.512613} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.873026] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 604.873026] env[65121]: INFO nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Took 12.86 seconds to spawn the instance on the hypervisor. [ 604.873026] env[65121]: DEBUG nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 604.874388] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357942dc-0691-461c-a66e-7ea9c8588869 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.881438] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105899, 'name': Rename_Task, 'duration_secs': 0.174416} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.886105] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 604.887588] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5508f1e7-0536-4c1e-967e-e5a8d152821c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.901320] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 604.901320] env[65121]: value = "task-5105904" [ 604.901320] env[65121]: _type = "Task" [ 604.901320] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.916040] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.976367] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105897, 'name': CreateVM_Task, 'duration_secs': 0.554275} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.976563] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 604.977164] env[65121]: WARNING neutronclient.v2_0.client [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.978032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.978032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 604.978032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 604.978195] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f550af88-0f90-4aff-8afe-3fd7d0e3517e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.985441] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 604.985441] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d59559-8601-d064-9951-ce605e3584ae" [ 604.985441] env[65121]: _type = "Task" [ 604.985441] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.002208] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d59559-8601-d064-9951-ce605e3584ae, 'name': SearchDatastore_Task, 'duration_secs': 0.010601} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.002208] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 605.002208] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 605.002208] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.002405] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 605.002405] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 605.002405] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97981de9-4cb2-49fb-b68e-a15c320f0c99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.022285] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 605.022509] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 605.023354] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c1f6959-d357-4162-a00a-91bc2d0db56a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.033032] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 605.033032] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b8f0df-2d51-2d18-d17b-9f89eb63a864" [ 605.033032] env[65121]: _type = "Task" [ 605.033032] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.044508] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b8f0df-2d51-2d18-d17b-9f89eb63a864, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.082816] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105902, 'name': CreateVM_Task, 'duration_secs': 0.382374} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.085286] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 605.086204] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d67e499-4f91-44eb-bdda-af82674a856f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.089364] env[65121]: WARNING neutronclient.v2_0.client [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 605.090350] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.090350] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 605.090350] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 605.090846] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-244b46c3-f302-4c20-ae55-eafc1766894b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.096658] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 605.096658] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b1002-350c-5770-75d9-e3e8d0fb4278" [ 605.096658] env[65121]: _type = "Task" [ 605.096658] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.102970] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d1050d-4d7c-4dae-963b-4422f4c3537e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.112512] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524b1002-350c-5770-75d9-e3e8d0fb4278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.145367] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71220a42-dcf1-456a-8067-8faf758eee1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.154697] env[65121]: DEBUG oslo_vmware.api [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105903, 'name': PowerOnVM_Task, 'duration_secs': 0.498957} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.157073] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 605.157363] env[65121]: INFO nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Took 8.59 seconds to spawn the instance on the hypervisor. [ 605.157545] env[65121]: DEBUG nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 605.158447] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210928cd-886a-48fb-9ae2-bb48a274d476 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.162843] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80aa63a-482d-4b53-82f5-573ca13a339e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.184581] env[65121]: DEBUG nova.compute.provider_tree [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.404300] env[65121]: INFO nova.compute.manager [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Took 17.65 seconds to build instance. [ 605.415117] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105904, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.540915] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b8f0df-2d51-2d18-d17b-9f89eb63a864, 'name': SearchDatastore_Task, 'duration_secs': 0.012272} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.541819] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f09b8d9f-4871-440b-a518-ed7bd2e1cdc6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.547484] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 605.547484] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528e67ed-482c-e756-2718-c01d45b4e8fe" [ 605.547484] env[65121]: _type = "Task" [ 605.547484] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.556270] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528e67ed-482c-e756-2718-c01d45b4e8fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.609025] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524b1002-350c-5770-75d9-e3e8d0fb4278, 'name': SearchDatastore_Task, 'duration_secs': 0.022388} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.609359] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 605.609584] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 605.609785] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.690781] env[65121]: DEBUG nova.scheduler.client.report [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 605.695712] env[65121]: INFO nova.compute.manager [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Took 14.62 seconds to build instance. [ 605.746344] env[65121]: DEBUG nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 605.778249] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 605.778747] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 605.778893] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 605.779048] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 605.779191] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 605.779329] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 605.779577] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.779730] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 605.779888] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 605.780057] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 605.780301] env[65121]: DEBUG nova.virt.hardware [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 605.781321] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3c6d00-9197-4576-a1ca-185083214260 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.791202] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbe2eb1-7c8e-4aa4-bf65-a44e729592dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.911844] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f508805e-5694-4d1a-b091-8ee241bcea78 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.173s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 605.922029] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105904, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.059883] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528e67ed-482c-e756-2718-c01d45b4e8fe, 'name': SearchDatastore_Task, 'duration_secs': 0.019816} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.059883] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 606.059883] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 0151ed85-bba4-443a-b18f-58272c973e1e/0151ed85-bba4-443a-b18f-58272c973e1e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 606.060148] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.060236] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 606.061307] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db7ff24c-3d8a-421e-a383-28013c537a80 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.064327] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a429d7f-2b4f-42d2-9fb6-180f82a59901 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.072396] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 606.072396] env[65121]: value = "task-5105905" [ 606.072396] env[65121]: _type = "Task" [ 606.072396] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.076771] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 606.076952] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 606.078017] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fda3486-7aff-46f3-a1ae-330dfce4f91a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.083621] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.087689] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 606.087689] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c3d23f-9b22-e56e-6ec4-e780d6e7706c" [ 606.087689] env[65121]: _type = "Task" [ 606.087689] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.096228] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c3d23f-9b22-e56e-6ec4-e780d6e7706c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.198575] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.198575] env[65121]: DEBUG nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 606.202028] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c6bc41-c329-40c1-9a97-6cf27cf81746 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "20396b81-2041-4bf2-85c4-8ea66c838ec4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.153s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.202140] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.898s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 606.203619] env[65121]: INFO nova.compute.claims [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.417653] env[65121]: DEBUG nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 606.426648] env[65121]: DEBUG nova.network.neutron [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Successfully updated port: 7aef5809-0a17-4d3e-b553-93f337e59e89 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 606.436413] env[65121]: DEBUG oslo_vmware.api [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5105904, 'name': PowerOnVM_Task, 'duration_secs': 1.156193} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.437204] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 606.437773] env[65121]: INFO nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Took 12.16 seconds to spawn the instance on the hypervisor. [ 606.437773] env[65121]: DEBUG nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 606.438615] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b17300a-3111-4a0c-b36e-95d17aa64d0d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.443116] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.443548] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.582649] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105905, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.599881] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c3d23f-9b22-e56e-6ec4-e780d6e7706c, 'name': SearchDatastore_Task, 'duration_secs': 0.013095} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.599881] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1c8ff55-c016-483c-8fbb-3af078af87c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.606228] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 606.606228] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52411652-a9c5-a10a-abc9-ad68bf99a22d" [ 606.606228] env[65121]: _type = "Task" [ 606.606228] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.615765] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52411652-a9c5-a10a-abc9-ad68bf99a22d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.713093] env[65121]: DEBUG nova.compute.utils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 606.715395] env[65121]: DEBUG nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 606.718752] env[65121]: DEBUG nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 606.718921] env[65121]: DEBUG nova.network.neutron [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 606.719164] env[65121]: WARNING neutronclient.v2_0.client [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 606.719888] env[65121]: WARNING neutronclient.v2_0.client [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 606.720268] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.720671] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.730316] env[65121]: DEBUG nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 606.938413] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "refresh_cache-2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.938413] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "refresh_cache-2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.938413] env[65121]: DEBUG nova.network.neutron [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 606.958919] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 606.971368] env[65121]: INFO nova.compute.manager [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Took 16.97 seconds to build instance. [ 607.083711] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.888339} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.085317] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 0151ed85-bba4-443a-b18f-58272c973e1e/0151ed85-bba4-443a-b18f-58272c973e1e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 607.085502] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 607.085825] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-48a616d2-9b72-41d9-8b49-bcdd1406c16e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.095047] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 607.095047] env[65121]: value = "task-5105906" [ 607.095047] env[65121]: _type = "Task" [ 607.095047] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.107164] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105906, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.117248] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52411652-a9c5-a10a-abc9-ad68bf99a22d, 'name': SearchDatastore_Task, 'duration_secs': 0.070859} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.117248] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 607.117248] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] a43879c7-7378-4700-9581-8daabdafe5b2/a43879c7-7378-4700-9581-8daabdafe5b2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 607.117971] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9d85f08-f978-4b6e-894e-9a2d5c3fb1b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.129933] env[65121]: DEBUG nova.policy [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfa7595ebb584044a2d0e8f3e469e9b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e92570886f544d9ea9fe0947e8e93d9c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 607.133606] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 607.133606] env[65121]: value = "task-5105907" [ 607.133606] env[65121]: _type = "Task" [ 607.133606] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.147810] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.256416] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 607.283557] env[65121]: WARNING neutronclient.v2_0.client [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 607.283557] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.283557] env[65121]: WARNING openstack [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.441974] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.448549] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.474575] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fb589d5-f428-49a6-b1a0-e0cab1540813 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "db737254-91fd-44bc-bf01-c864227cdf07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.490s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 607.562375] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0c7fc2-f128-4d7e-be3f-0e6e58313f40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.575263] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d0e3f8-8bcb-4062-b527-2c02e63ffa18 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.617797] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1793a4-ebc3-452c-bd2b-2d17fb475898 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.626987] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105906, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101215} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.630072] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 607.630934] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b606a6a-41e4-4887-9eef-cc2ea65f6761 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.635231] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a58378-9132-4a7b-bf36-094dd1352a9c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.667824] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 0151ed85-bba4-443a-b18f-58272c973e1e/0151ed85-bba4-443a-b18f-58272c973e1e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 607.678064] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-861cb38c-4145-4666-9b11-44afc14cae8b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.693684] env[65121]: DEBUG nova.compute.provider_tree [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.706080] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 607.706080] env[65121]: value = "task-5105908" [ 607.706080] env[65121]: _type = "Task" [ 607.706080] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.708705] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105907, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.722472] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105908, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.741876] env[65121]: DEBUG nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 607.775575] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 607.775848] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 607.776041] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 607.776271] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 607.776444] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 607.776600] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 607.776834] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.777056] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 607.777245] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 607.777442] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 607.777655] env[65121]: DEBUG nova.virt.hardware [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 607.778659] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06dc042-8f40-4a1d-b309-cb7262c8d097 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.791208] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a27bcf2-42cb-4675-9325-d845bd254fb8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.810125] env[65121]: DEBUG nova.network.neutron [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Successfully created port: f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 607.983372] env[65121]: DEBUG nova.network.neutron [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 607.988839] env[65121]: DEBUG nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 608.111393] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "46f89cfb-14aa-483d-aa8a-bc91356272ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 608.111559] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 608.150858] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105907, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.662283} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.151205] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] a43879c7-7378-4700-9581-8daabdafe5b2/a43879c7-7378-4700-9581-8daabdafe5b2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 608.151659] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 608.151750] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5b11723-d269-4565-8cce-06ce020af394 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.161329] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 608.161329] env[65121]: value = "task-5105909" [ 608.161329] env[65121]: _type = "Task" [ 608.161329] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.170976] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.199051] env[65121]: DEBUG nova.scheduler.client.report [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 608.219861] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105908, 'name': ReconfigVM_Task, 'duration_secs': 0.363736} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.220323] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 0151ed85-bba4-443a-b18f-58272c973e1e/0151ed85-bba4-443a-b18f-58272c973e1e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 608.221092] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afb6801e-a449-49e1-88c8-f5ee75ccc50b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.228798] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 608.228798] env[65121]: value = "task-5105910" [ 608.228798] env[65121]: _type = "Task" [ 608.228798] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.243460] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105910, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.513588] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 608.673023] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085498} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.673280] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 608.674255] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037a04d8-5e71-46f9-989f-5f076c0b5fdd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.697565] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] a43879c7-7378-4700-9581-8daabdafe5b2/a43879c7-7378-4700-9581-8daabdafe5b2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 608.698045] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f951a9c5-9181-44da-a3a1-cf397965c507 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.713287] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 608.713805] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 608.716477] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.360s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 608.717727] env[65121]: INFO nova.compute.claims [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.726644] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 608.726644] env[65121]: value = "task-5105911" [ 608.726644] env[65121]: _type = "Task" [ 608.726644] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.741382] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105911, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.741627] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105910, 'name': Rename_Task, 'duration_secs': 0.164329} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.742012] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 608.742153] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67ba94eb-0ff1-4852-b08e-f3bfbd9018d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.751029] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 608.751029] env[65121]: value = "task-5105912" [ 608.751029] env[65121]: _type = "Task" [ 608.751029] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.759837] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105912, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.106720] env[65121]: DEBUG nova.network.neutron [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Updated VIF entry in instance network info cache for port a059ff79-fc05-40c7-9def-3f3c476870b8. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 609.107084] env[65121]: DEBUG nova.network.neutron [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Updating instance_info_cache with network_info: [{"id": "a059ff79-fc05-40c7-9def-3f3c476870b8", "address": "fa:16:3e:0a:65:44", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.124", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa059ff79-fc", "ovs_interfaceid": "a059ff79-fc05-40c7-9def-3f3c476870b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 609.123707] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.124185] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.225367] env[65121]: DEBUG nova.compute.utils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 609.231485] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 609.231485] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 609.231485] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.231485] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.231485] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.231658] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.259374] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105911, 'name': ReconfigVM_Task, 'duration_secs': 0.516231} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.264773] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Reconfigured VM instance instance-00000005 to attach disk [datastore1] a43879c7-7378-4700-9581-8daabdafe5b2/a43879c7-7378-4700-9581-8daabdafe5b2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 609.266934] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52e33fd8-4a5d-4ff5-a03d-cd2ae2c78261 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.280020] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105912, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.280020] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 609.280020] env[65121]: value = "task-5105913" [ 609.280020] env[65121]: _type = "Task" [ 609.280020] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.291934] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105913, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.610395] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca927e9b-d9a5-4b2b-93f9-4eded7ae2267 req-f0c17f0c-aaa0-4df7-b9e3-83cd52fc230e service nova] Releasing lock "refresh_cache-db737254-91fd-44bc-bf01-c864227cdf07" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.618962] env[65121]: DEBUG nova.policy [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea485ec50b264dfeb9e93ff0ae5a7544', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f0b5fd1c88e44d3952279db674447a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 609.720906] env[65121]: WARNING neutronclient.v2_0.client [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.721339] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.721707] env[65121]: WARNING openstack [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.730180] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 609.773669] env[65121]: DEBUG nova.network.neutron [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Successfully updated port: f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 609.779674] env[65121]: DEBUG oslo_vmware.api [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105912, 'name': PowerOnVM_Task, 'duration_secs': 0.653894} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.784492] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 609.784492] env[65121]: INFO nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Took 10.89 seconds to spawn the instance on the hypervisor. [ 609.784492] env[65121]: DEBUG nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 609.785853] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e99801-774a-4838-97fb-583c87ff617c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.799137] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105913, 'name': Rename_Task, 'duration_secs': 0.219244} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.804862] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 609.805879] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f822371-e6b8-4dfa-96fb-b33401254bb7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.818016] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 609.818016] env[65121]: value = "task-5105914" [ 609.818016] env[65121]: _type = "Task" [ 609.818016] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.833532] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.049655] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f3bd34-914b-43af-bdb7-103f243c62d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.063675] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d03622-e55f-4cc5-8050-6d5f1f3acbfc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.096819] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f97dc29-12f9-4a85-ab3b-226701037cd0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.102855] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Successfully created port: 1542237c-e4ab-42a7-a285-090f397a4a70 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 610.109128] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4146a9f-2d5b-42f6-bb12-454fee629e5d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.123989] env[65121]: DEBUG nova.compute.provider_tree [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.281104] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.281386] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.281755] env[65121]: DEBUG nova.network.neutron [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 610.314702] env[65121]: INFO nova.compute.manager [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Took 17.98 seconds to build instance. [ 610.329476] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105914, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.629052] env[65121]: DEBUG nova.scheduler.client.report [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 610.741217] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 610.778621] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 610.778898] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 610.779204] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 610.779387] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 610.779528] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 610.779776] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 610.779873] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.779996] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 610.780171] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 610.780319] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 610.780480] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 610.781966] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bac10ac-d64d-41a5-bdea-3fe64f020714 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.785632] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.786470] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.801733] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019afb4a-9b75-46dd-853c-30910547d367 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.823098] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ebe1b35-faf5-4e49-afd8-099b33645004 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "0151ed85-bba4-443a-b18f-58272c973e1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.499s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 610.840065] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105914, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.135823] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 611.136370] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 611.141321] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.517s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 611.339518] env[65121]: DEBUG nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 611.340317] env[65121]: DEBUG oslo_vmware.api [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5105914, 'name': PowerOnVM_Task, 'duration_secs': 1.058566} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.340576] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 611.340764] env[65121]: INFO nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Took 10.04 seconds to spawn the instance on the hypervisor. [ 611.340926] env[65121]: DEBUG nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 611.342116] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddef5ce9-99b7-4cb8-836f-26264817b31a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.491652] env[65121]: DEBUG nova.network.neutron [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Updating instance_info_cache with network_info: [{"id": "7aef5809-0a17-4d3e-b553-93f337e59e89", "address": "fa:16:3e:81:87:46", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aef5809-0a", "ovs_interfaceid": "7aef5809-0a17-4d3e-b553-93f337e59e89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 611.549697] env[65121]: DEBUG nova.network.neutron [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 611.647167] env[65121]: DEBUG nova.compute.utils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 611.655745] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 611.655870] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 611.657191] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.657191] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.657531] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.658000] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.864977] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.867275] env[65121]: INFO nova.compute.manager [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Took 19.42 seconds to build instance. [ 611.926386] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Successfully updated port: 1542237c-e4ab-42a7-a285-090f397a4a70 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 611.962185] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.962605] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.995564] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "refresh_cache-2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 611.995564] env[65121]: DEBUG nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Instance network_info: |[{"id": "7aef5809-0a17-4d3e-b553-93f337e59e89", "address": "fa:16:3e:81:87:46", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aef5809-0a", "ovs_interfaceid": "7aef5809-0a17-4d3e-b553-93f337e59e89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 611.995726] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:87:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7aef5809-0a17-4d3e-b553-93f337e59e89', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 612.004244] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating folder: Project (40aee06073b04c25bc7b38fbd09856e9). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.004344] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59f670c0-b58e-4f44-bd98-e1b4161a7263 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.016523] env[65121]: DEBUG nova.policy [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea485ec50b264dfeb9e93ff0ae5a7544', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f0b5fd1c88e44d3952279db674447a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 612.020210] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created folder: Project (40aee06073b04c25bc7b38fbd09856e9) in parent group-v993268. [ 612.020411] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating folder: Instances. Parent ref: group-v993284. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.020691] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bc2f5dd-4e20-49ea-b983-5fd1e229aff7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.035549] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created folder: Instances in parent group-v993284. [ 612.035549] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 612.035855] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 612.040747] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-303a5449-8463-4cc0-b88b-8a41340319e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.067977] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 612.067977] env[65121]: value = "task-5105917" [ 612.067977] env[65121]: _type = "Task" [ 612.067977] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.076587] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105917, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.124094] env[65121]: WARNING neutronclient.v2_0.client [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.124832] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.125202] env[65121]: WARNING openstack [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.157071] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 612.208298] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 5f6e4f46-0745-42c0-a779-6cffb60e21ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.208468] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance db737254-91fd-44bc-bf01-c864227cdf07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.208607] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 20396b81-2041-4bf2-85c4-8ea66c838ec4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.208734] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 0151ed85-bba4-443a-b18f-58272c973e1e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.208858] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance a43879c7-7378-4700-9581-8daabdafe5b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.209017] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.209147] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance aa9b6708-c53c-4117-9b75-9d506f393395 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.209264] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance bad21c0b-6534-46b5-97a3-a26e5ecb5c96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.209372] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 5e915c8e-150a-4bfd-8005-a7a1a119057d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 612.373095] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d161c50d-9485-4f3e-ab8a-a4fd5f4be2ae tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "a43879c7-7378-4700-9581-8daabdafe5b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.934s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.405411] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Successfully created port: 062568cf-fd2f-407b-8cf8-da76b66e5c30 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 612.421388] env[65121]: DEBUG nova.compute.manager [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Received event network-changed-cf0d4936-618c-45a3-af9c-3f745b2cf316 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 612.421388] env[65121]: DEBUG nova.compute.manager [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Refreshing instance network info cache due to event network-changed-cf0d4936-618c-45a3-af9c-3f745b2cf316. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 612.421388] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Acquiring lock "refresh_cache-0151ed85-bba4-443a-b18f-58272c973e1e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.421388] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Acquired lock "refresh_cache-0151ed85-bba4-443a-b18f-58272c973e1e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 612.421388] env[65121]: DEBUG nova.network.neutron [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Refreshing network info cache for port cf0d4936-618c-45a3-af9c-3f745b2cf316 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 612.429438] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "refresh_cache-bad21c0b-6534-46b5-97a3-a26e5ecb5c96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.429747] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "refresh_cache-bad21c0b-6534-46b5-97a3-a26e5ecb5c96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 612.433452] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 612.482519] env[65121]: DEBUG nova.compute.manager [req-407c59e0-37bb-44f0-93de-3fb1b4bf9ddc req-312d031f-fdf6-4254-be3a-dc9243ca1d35 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Received event network-vif-plugged-7aef5809-0a17-4d3e-b553-93f337e59e89 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 612.482672] env[65121]: DEBUG oslo_concurrency.lockutils [req-407c59e0-37bb-44f0-93de-3fb1b4bf9ddc req-312d031f-fdf6-4254-be3a-dc9243ca1d35 service nova] Acquiring lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.482859] env[65121]: DEBUG oslo_concurrency.lockutils [req-407c59e0-37bb-44f0-93de-3fb1b4bf9ddc req-312d031f-fdf6-4254-be3a-dc9243ca1d35 service nova] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.483096] env[65121]: DEBUG oslo_concurrency.lockutils [req-407c59e0-37bb-44f0-93de-3fb1b4bf9ddc req-312d031f-fdf6-4254-be3a-dc9243ca1d35 service nova] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.483226] env[65121]: DEBUG nova.compute.manager [req-407c59e0-37bb-44f0-93de-3fb1b4bf9ddc req-312d031f-fdf6-4254-be3a-dc9243ca1d35 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] No waiting events found dispatching network-vif-plugged-7aef5809-0a17-4d3e-b553-93f337e59e89 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 612.483479] env[65121]: WARNING nova.compute.manager [req-407c59e0-37bb-44f0-93de-3fb1b4bf9ddc req-312d031f-fdf6-4254-be3a-dc9243ca1d35 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Received unexpected event network-vif-plugged-7aef5809-0a17-4d3e-b553-93f337e59e89 for instance with vm_state building and task_state spawning. [ 612.505916] env[65121]: DEBUG nova.network.neutron [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updating instance_info_cache with network_info: [{"id": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "address": "fa:16:3e:47:bb:ea", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119a70b-b1", "ovs_interfaceid": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 612.579043] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105917, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.714926] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance cc002961-d742-4255-88c7-f0a5a39424b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 612.875637] env[65121]: DEBUG nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 612.925180] env[65121]: WARNING neutronclient.v2_0.client [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.926013] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.926013] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.942032] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.942440] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.012204] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Releasing lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 613.012691] env[65121]: DEBUG nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Instance network_info: |[{"id": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "address": "fa:16:3e:47:bb:ea", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119a70b-b1", "ovs_interfaceid": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 613.013156] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:bb:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f119a70b-b122-4875-b1a3-5c4c7a234c29', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 613.026991] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Creating folder: Project (e92570886f544d9ea9fe0947e8e93d9c). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 613.027909] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-445757d9-f675-46ce-aaed-8c46ba3b5cf7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.043196] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Created folder: Project (e92570886f544d9ea9fe0947e8e93d9c) in parent group-v993268. [ 613.044292] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Creating folder: Instances. Parent ref: group-v993287. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 613.044862] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d802cb2-c10c-426f-811d-9774bc1d4f3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.060574] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Created folder: Instances in parent group-v993287. [ 613.061330] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 613.062145] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 613.062712] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5bcb68d-dd40-4af5-87a8-02848fafbcf1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.095251] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105917, 'name': CreateVM_Task, 'duration_secs': 0.51528} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.097622] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 613.097622] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 613.097622] env[65121]: value = "task-5105920" [ 613.097622] env[65121]: _type = "Task" [ 613.097622] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.097622] env[65121]: WARNING neutronclient.v2_0.client [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.098685] env[65121]: DEBUG oslo_vmware.service [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83572aab-9c56-4413-8150-7f916f3c6224 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.112214] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.112355] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 613.113284] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 613.113284] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e564d8f4-fa79-4e56-80e2-c28e6a4f76d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.119128] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 613.119128] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dd079b-100e-94ef-afbb-1883b68cde5e" [ 613.119128] env[65121]: _type = "Task" [ 613.119128] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.132036] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dd079b-100e-94ef-afbb-1883b68cde5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.147310] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 613.169696] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 613.205270] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 613.205571] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 613.205805] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 613.205946] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 613.206219] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 613.206519] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 613.206519] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.207480] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 613.207480] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 613.207480] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 613.207480] env[65121]: DEBUG nova.virt.hardware [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 613.209271] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33368ba-d527-4bc8-8d8c-fc2ffa3fb8eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.220486] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f1a5fdce-a940-43d4-83d7-716786c9ac34 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 613.228626] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7537bf42-2ea7-4f76-a690-34fcd67a2a5a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.408161] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.612341] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105920, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.630222] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 613.630482] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.630719] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.630856] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 613.631036] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.631299] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6a0e2c4-f2c3-4a20-8049-c33153f576d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.640796] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.640979] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 613.641786] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a75170-c796-4086-b587-5369f958e0d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.648935] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec66ebe-694e-4eb1-9826-08b386aa46c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.655182] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 613.655182] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c6b1ad-1dda-c0e2-5e9d-3f560408bf1f" [ 613.655182] env[65121]: _type = "Task" [ 613.655182] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.663514] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c6b1ad-1dda-c0e2-5e9d-3f560408bf1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.732640] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 6887bb66-7186-4d58-be59-d0c53c3221c6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 613.790932] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.791387] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.806101] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.806534] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.960642] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.961075] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.961418] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.992058] env[65121]: WARNING neutronclient.v2_0.client [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.992739] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.993099] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.009193] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Successfully updated port: 062568cf-fd2f-407b-8cf8-da76b66e5c30 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 614.093810] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Updating instance_info_cache with network_info: [{"id": "1542237c-e4ab-42a7-a285-090f397a4a70", "address": "fa:16:3e:8f:18:ce", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1542237c-e4", "ovs_interfaceid": "1542237c-e4ab-42a7-a285-090f397a4a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 614.116973] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105920, 'name': CreateVM_Task, 'duration_secs': 0.538436} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.117257] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 614.117807] env[65121]: WARNING neutronclient.v2_0.client [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.118259] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.118474] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 614.118811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 614.119111] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d8a88bf-5179-4c5b-98aa-0b6dc5ab00ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.124978] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 614.124978] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5209c4c2-c230-548c-1b86-bca3c0e9a21a" [ 614.124978] env[65121]: _type = "Task" [ 614.124978] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.127825] env[65121]: DEBUG nova.network.neutron [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Updated VIF entry in instance network info cache for port cf0d4936-618c-45a3-af9c-3f745b2cf316. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 614.127825] env[65121]: DEBUG nova.network.neutron [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Updating instance_info_cache with network_info: [{"id": "cf0d4936-618c-45a3-af9c-3f745b2cf316", "address": "fa:16:3e:88:13:d9", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf0d4936-61", "ovs_interfaceid": "cf0d4936-618c-45a3-af9c-3f745b2cf316", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 614.140135] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5209c4c2-c230-548c-1b86-bca3c0e9a21a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.166266] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Preparing fetch location {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 614.166539] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating directory with path [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 614.166772] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91c021dc-c5a6-49e0-be71-4cde47898baa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.190866] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created directory with path [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 614.191105] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Fetch image to [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 614.191330] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Downloading image file data 3ed5d656-a616-4609-8bb7-b02beb3ac3df to [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk on the data store datastore2 {{(pid=65121) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 614.192151] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23bb585-c973-4980-8aa0-a014c53c0874 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.200918] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52675dd7-b18c-43ab-b6a0-9543345235fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.212288] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18616ca9-e7a6-488b-97f7-bbfc76af2e43 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.243913] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c4096314-270a-4270-9e1d-5ace8ddbd286 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 614.245927] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62564513-1f95-4e5b-8048-3485044620ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.254284] env[65121]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-a0034643-fa0c-431b-bade-3bb7e5df7ff6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.282799] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Downloading image file data 3ed5d656-a616-4609-8bb7-b02beb3ac3df to the data store datastore2 {{(pid=65121) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 614.333877] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65121) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 614.514024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "refresh_cache-5e915c8e-150a-4bfd-8005-a7a1a119057d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.514024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "refresh_cache-5e915c8e-150a-4bfd-8005-a7a1a119057d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 614.514024] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 614.597040] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "refresh_cache-bad21c0b-6534-46b5-97a3-a26e5ecb5c96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 614.597690] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Instance network_info: |[{"id": "1542237c-e4ab-42a7-a285-090f397a4a70", "address": "fa:16:3e:8f:18:ce", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1542237c-e4", "ovs_interfaceid": "1542237c-e4ab-42a7-a285-090f397a4a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 614.598438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:18:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1542237c-e4ab-42a7-a285-090f397a4a70', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 614.609764] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Creating folder: Project (3f0b5fd1c88e44d3952279db674447a6). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 614.609764] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-641cc3fb-8c40-4492-a3f6-dd87e4098a94 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.627916] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Created folder: Project (3f0b5fd1c88e44d3952279db674447a6) in parent group-v993268. [ 614.629052] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Creating folder: Instances. Parent ref: group-v993290. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 614.629052] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a825858a-8280-4fea-924c-a279b381482a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.634637] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Releasing lock "refresh_cache-0151ed85-bba4-443a-b18f-58272c973e1e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 614.634847] env[65121]: DEBUG nova.compute.manager [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Received event network-vif-plugged-e6d39f1f-d115-4218-a28a-e2cd0d05868d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 614.635175] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Acquiring lock "a43879c7-7378-4700-9581-8daabdafe5b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.635252] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Lock "a43879c7-7378-4700-9581-8daabdafe5b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 614.635578] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Lock "a43879c7-7378-4700-9581-8daabdafe5b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 614.635578] env[65121]: DEBUG nova.compute.manager [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] No waiting events found dispatching network-vif-plugged-e6d39f1f-d115-4218-a28a-e2cd0d05868d {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 614.635705] env[65121]: WARNING nova.compute.manager [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Received unexpected event network-vif-plugged-e6d39f1f-d115-4218-a28a-e2cd0d05868d for instance with vm_state active and task_state None. [ 614.635851] env[65121]: DEBUG nova.compute.manager [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Received event network-changed-e6d39f1f-d115-4218-a28a-e2cd0d05868d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 614.636144] env[65121]: DEBUG nova.compute.manager [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Refreshing instance network info cache due to event network-changed-e6d39f1f-d115-4218-a28a-e2cd0d05868d. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 614.636252] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Acquiring lock "refresh_cache-a43879c7-7378-4700-9581-8daabdafe5b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.636368] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Acquired lock "refresh_cache-a43879c7-7378-4700-9581-8daabdafe5b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 614.637405] env[65121]: DEBUG nova.network.neutron [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Refreshing network info cache for port e6d39f1f-d115-4218-a28a-e2cd0d05868d {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 614.646291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 614.646577] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 614.646811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.647043] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Created folder: Instances in parent group-v993290. [ 614.647311] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 614.648104] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 614.648323] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82517742-b19f-430c-858e-f1448d811519 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.681861] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 614.681861] env[65121]: value = "task-5105923" [ 614.681861] env[65121]: _type = "Task" [ 614.681861] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.696228] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105923, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.750355] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 614.898174] env[65121]: INFO nova.compute.manager [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Rebuilding instance [ 614.958058] env[65121]: DEBUG nova.compute.manager [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 614.958911] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6441013-a2e3-48de-b239-054cdb0af33a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.016581] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.017303] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.131543] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 615.139798] env[65121]: WARNING neutronclient.v2_0.client [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.142278] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.142278] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.199120] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105923, 'name': CreateVM_Task, 'duration_secs': 0.462137} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.199963] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 615.199963] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.200818] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.200818] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 615.200818] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 615.201593] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f59feb4-fb18-4f78-a3a8-532d9cbcc94f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.207688] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 615.207688] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526923fa-613c-1a26-927d-aef47e8e61b4" [ 615.207688] env[65121]: _type = "Task" [ 615.207688] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.217439] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Acquiring lock "db737254-91fd-44bc-bf01-c864227cdf07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 615.217622] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Lock "db737254-91fd-44bc-bf01-c864227cdf07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.217894] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Acquiring lock "db737254-91fd-44bc-bf01-c864227cdf07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 615.218116] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Lock "db737254-91fd-44bc-bf01-c864227cdf07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.218320] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Lock "db737254-91fd-44bc-bf01-c864227cdf07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.223943] env[65121]: INFO nova.compute.manager [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Terminating instance [ 615.237346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 615.237346] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 615.237346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.250197] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.251643] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.258545] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 46f89cfb-14aa-483d-aa8a-bc91356272ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 615.258822] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 615.258996] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=100GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '9', 'num_vm_active': '4', 'num_task_None': '5', 'num_os_type_None': '9', 'num_proj_982f66705583488cb369f19160cc2ee5': '1', 'io_workload': '5', 'num_proj_a92a7a59dd5c44f9b925030ae81e344d': '1', 'num_proj_21027ead4f1042029f75e43c3e015eb8': '1', 'num_proj_1d44b5f8148445c8bbaa5fd09032c0c0': '1', 'num_vm_building': '5', 'num_task_spawning': '4', 'num_proj_018269ed0c644392aebdb56cf0523292': '1', 'num_proj_40aee06073b04c25bc7b38fbd09856e9': '1', 'num_proj_e92570886f544d9ea9fe0947e8e93d9c': '1', 'num_proj_3f0b5fd1c88e44d3952279db674447a6': '2'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 615.271300] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Completed reading data from the image iterator. {{(pid=65121) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 615.271509] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 615.348779] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Downloaded image file data 3ed5d656-a616-4609-8bb7-b02beb3ac3df to vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk on the data store datastore2 {{(pid=65121) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 615.350797] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Caching image {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 615.350797] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copying Virtual Disk [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk to [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 615.351033] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6001a862-9b95-465a-b34a-c1a0b4ccfce1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.360568] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 615.360568] env[65121]: value = "task-5105924" [ 615.360568] env[65121]: _type = "Task" [ 615.360568] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.365613] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.366630] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.367042] env[65121]: WARNING openstack [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.396795] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.550008] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41deecc8-edf4-4ad4-9264-6c70ca630e99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.558874] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd905b44-a74e-4bf0-b72a-b3e69d8b9d92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.593990] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58e77ad-be16-47b1-a6b2-1b55a1ff0223 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.602262] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72016ad-72b4-417e-8596-673074da0e3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.607336] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.607701] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.626438] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.737365] env[65121]: DEBUG nova.compute.manager [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 615.737365] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.737565] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbbe3ec-612a-4ae8-acbf-4f1a9d2f499d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.748716] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 615.748716] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72b642f5-9f35-4f44-9c67-92bf1422da3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.754274] env[65121]: DEBUG oslo_vmware.api [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Waiting for the task: (returnval){ [ 615.754274] env[65121]: value = "task-5105925" [ 615.754274] env[65121]: _type = "Task" [ 615.754274] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.765660] env[65121]: DEBUG oslo_vmware.api [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Task: {'id': task-5105925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.790591] env[65121]: DEBUG nova.network.neutron [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Updating instance_info_cache with network_info: [{"id": "062568cf-fd2f-407b-8cf8-da76b66e5c30", "address": "fa:16:3e:5c:d1:31", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap062568cf-fd", "ovs_interfaceid": "062568cf-fd2f-407b-8cf8-da76b66e5c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 615.874509] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105924, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.902621] env[65121]: WARNING neutronclient.v2_0.client [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.903628] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.907307] env[65121]: WARNING openstack [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.975499] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 615.975927] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2577db6-bd75-4e16-aeb3-fa76a9f39ee3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.984803] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 615.984803] env[65121]: value = "task-5105926" [ 615.984803] env[65121]: _type = "Task" [ 615.984803] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.995640] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.048699] env[65121]: DEBUG nova.network.neutron [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Updated VIF entry in instance network info cache for port e6d39f1f-d115-4218-a28a-e2cd0d05868d. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 616.049099] env[65121]: DEBUG nova.network.neutron [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Updating instance_info_cache with network_info: [{"id": "e6d39f1f-d115-4218-a28a-e2cd0d05868d", "address": "fa:16:3e:4b:6e:b3", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.158", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6d39f1f-d1", "ovs_interfaceid": "e6d39f1f-d115-4218-a28a-e2cd0d05868d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 616.131365] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 616.253601] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Received event network-vif-plugged-1542237c-e4ab-42a7-a285-090f397a4a70 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 616.253810] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Acquiring lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.254035] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.256686] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.003s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 616.256866] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] No waiting events found dispatching network-vif-plugged-1542237c-e4ab-42a7-a285-090f397a4a70 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 616.257054] env[65121]: WARNING nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Received unexpected event network-vif-plugged-1542237c-e4ab-42a7-a285-090f397a4a70 for instance with vm_state building and task_state spawning. [ 616.257189] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Received event network-changed-1542237c-e4ab-42a7-a285-090f397a4a70 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 616.257331] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Refreshing instance network info cache due to event network-changed-1542237c-e4ab-42a7-a285-090f397a4a70. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 616.257508] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Acquiring lock "refresh_cache-bad21c0b-6534-46b5-97a3-a26e5ecb5c96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.257632] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Acquired lock "refresh_cache-bad21c0b-6534-46b5-97a3-a26e5ecb5c96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 616.257778] env[65121]: DEBUG nova.network.neutron [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Refreshing network info cache for port 1542237c-e4ab-42a7-a285-090f397a4a70 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 616.272687] env[65121]: DEBUG oslo_vmware.api [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Task: {'id': task-5105925, 'name': PowerOffVM_Task, 'duration_secs': 0.2933} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.272933] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 616.273102] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 616.273370] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1ca331d-69cb-49b4-819a-bbfc59ea63d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.278851] env[65121]: DEBUG nova.compute.manager [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Received event network-changed-7aef5809-0a17-4d3e-b553-93f337e59e89 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 616.279038] env[65121]: DEBUG nova.compute.manager [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Refreshing instance network info cache due to event network-changed-7aef5809-0a17-4d3e-b553-93f337e59e89. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 616.279239] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Acquiring lock "refresh_cache-2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.279404] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Acquired lock "refresh_cache-2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 616.279601] env[65121]: DEBUG nova.network.neutron [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Refreshing network info cache for port 7aef5809-0a17-4d3e-b553-93f337e59e89 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 616.293130] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "refresh_cache-5e915c8e-150a-4bfd-8005-a7a1a119057d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 616.293621] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Instance network_info: |[{"id": "062568cf-fd2f-407b-8cf8-da76b66e5c30", "address": "fa:16:3e:5c:d1:31", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap062568cf-fd", "ovs_interfaceid": "062568cf-fd2f-407b-8cf8-da76b66e5c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 616.294316] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:d1:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '062568cf-fd2f-407b-8cf8-da76b66e5c30', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 616.304218] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 616.305026] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 616.305259] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aca6cd58-7c77-4262-9508-92faf46679f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.329132] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 616.329132] env[65121]: value = "task-5105928" [ 616.329132] env[65121]: _type = "Task" [ 616.329132] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.339166] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105928, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.374516] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105924, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.714111} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.374806] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copied Virtual Disk [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk to [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 616.375111] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df/tmp-sparse.vmdk {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 616.375262] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d0b7d03-f657-46de-a698-5b9ab487edd2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.382908] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 616.382908] env[65121]: value = "task-5105929" [ 616.382908] env[65121]: _type = "Task" [ 616.382908] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.392356] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.496098] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105926, 'name': PowerOffVM_Task, 'duration_secs': 0.162223} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.496734] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 616.496734] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 616.499607] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-003f128a-8449-496f-bd67-e91becef3615 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.520997] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 616.522645] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6ad98e2-d287-4e1c-949a-be5dc87888ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.556793] env[65121]: DEBUG oslo_concurrency.lockutils [req-5defa04d-77b4-4d38-bacb-86efff062ecb req-de314ad6-39a5-4a7b-a9f7-2de5b325b35e service nova] Releasing lock "refresh_cache-a43879c7-7378-4700-9581-8daabdafe5b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 616.561116] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 616.561441] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 616.561656] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleting the datastore file [datastore1] 20396b81-2041-4bf2-85c4-8ea66c838ec4 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 616.562502] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f106d69-cfbd-49cc-aa4f-57c29ac20452 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.571087] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 616.571087] env[65121]: value = "task-5105931" [ 616.571087] env[65121]: _type = "Task" [ 616.571087] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.580035] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.637110] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 616.637359] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.497s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 616.637624] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.997s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.639335] env[65121]: INFO nova.compute.claims [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.689372] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 616.689649] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 616.689757] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Deleting the datastore file [datastore1] db737254-91fd-44bc-bf01-c864227cdf07 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 616.690033] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f934e57d-004c-420f-8572-4a4571e96eec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.703605] env[65121]: DEBUG oslo_vmware.api [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Waiting for the task: (returnval){ [ 616.703605] env[65121]: value = "task-5105932" [ 616.703605] env[65121]: _type = "Task" [ 616.703605] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.713435] env[65121]: DEBUG oslo_vmware.api [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Task: {'id': task-5105932, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.761031] env[65121]: WARNING neutronclient.v2_0.client [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.762019] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.762151] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.784030] env[65121]: WARNING neutronclient.v2_0.client [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.785575] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.785575] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.839759] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105928, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.859489] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.859748] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.902713] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02783} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.903452] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 616.903452] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Moving file from [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241/3ed5d656-a616-4609-8bb7-b02beb3ac3df to [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df. {{(pid=65121) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 616.903703] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-38d917a5-3beb-40ec-abb7-0bb01f0fbce8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.914436] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 616.914436] env[65121]: value = "task-5105933" [ 616.914436] env[65121]: _type = "Task" [ 616.914436] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.930018] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105933, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.935825] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.936234] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.944871] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.945325] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.017123] env[65121]: DEBUG nova.compute.manager [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 617.043202] env[65121]: WARNING neutronclient.v2_0.client [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.043887] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.044206] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.054769] env[65121]: WARNING neutronclient.v2_0.client [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.054769] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.055337] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.084688] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10464} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.084688] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 617.084688] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 617.084688] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 617.143659] env[65121]: DEBUG nova.network.neutron [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Updated VIF entry in instance network info cache for port 7aef5809-0a17-4d3e-b553-93f337e59e89. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 617.143969] env[65121]: DEBUG nova.network.neutron [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Updating instance_info_cache with network_info: [{"id": "7aef5809-0a17-4d3e-b553-93f337e59e89", "address": "fa:16:3e:81:87:46", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aef5809-0a", "ovs_interfaceid": "7aef5809-0a17-4d3e-b553-93f337e59e89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 617.177966] env[65121]: DEBUG nova.network.neutron [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Updated VIF entry in instance network info cache for port 1542237c-e4ab-42a7-a285-090f397a4a70. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 617.178342] env[65121]: DEBUG nova.network.neutron [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Updating instance_info_cache with network_info: [{"id": "1542237c-e4ab-42a7-a285-090f397a4a70", "address": "fa:16:3e:8f:18:ce", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1542237c-e4", "ovs_interfaceid": "1542237c-e4ab-42a7-a285-090f397a4a70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 617.214565] env[65121]: DEBUG oslo_vmware.api [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Task: {'id': task-5105932, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222447} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.214807] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 617.215410] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 617.216397] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 617.216397] env[65121]: INFO nova.compute.manager [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Took 1.48 seconds to destroy the instance on the hypervisor. [ 617.217409] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 617.217409] env[65121]: DEBUG nova.compute.manager [-] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 617.217409] env[65121]: DEBUG nova.network.neutron [-] [instance: db737254-91fd-44bc-bf01-c864227cdf07] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 617.217409] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.217752] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.218009] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.290510] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.341691] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105928, 'name': CreateVM_Task, 'duration_secs': 0.639185} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.341879] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 617.342383] env[65121]: WARNING neutronclient.v2_0.client [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.342731] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.342876] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.343361] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 617.343477] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d98cb381-5b21-47f8-afef-da0569dba128 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.349552] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 617.349552] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5217fc3e-44ff-fe7b-3bf6-8ee1abbd2818" [ 617.349552] env[65121]: _type = "Task" [ 617.349552] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.359043] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5217fc3e-44ff-fe7b-3bf6-8ee1abbd2818, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.424458] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105933, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.048656} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.426474] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] File moved {{(pid=65121) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 617.426562] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Cleaning up location [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 617.426864] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore2] vmware_temp/341c2c4d-621b-4556-8a3e-730bd80e2241 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 617.426975] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3937295a-f543-4b7d-9985-cd5e69f40418 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.435900] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 617.435900] env[65121]: value = "task-5105934" [ 617.435900] env[65121]: _type = "Task" [ 617.435900] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.446744] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105934, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.537276] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.569989] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "0151ed85-bba4-443a-b18f-58272c973e1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.569989] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "0151ed85-bba4-443a-b18f-58272c973e1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.569989] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "0151ed85-bba4-443a-b18f-58272c973e1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.569989] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "0151ed85-bba4-443a-b18f-58272c973e1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.570240] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "0151ed85-bba4-443a-b18f-58272c973e1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.571835] env[65121]: INFO nova.compute.manager [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Terminating instance [ 617.648293] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Releasing lock "refresh_cache-2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.648567] env[65121]: DEBUG nova.compute.manager [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received event network-vif-plugged-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 617.648749] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Acquiring lock "aa9b6708-c53c-4117-9b75-9d506f393395-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.648965] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Lock "aa9b6708-c53c-4117-9b75-9d506f393395-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.649130] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Lock "aa9b6708-c53c-4117-9b75-9d506f393395-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.649317] env[65121]: DEBUG nova.compute.manager [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] No waiting events found dispatching network-vif-plugged-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 617.649558] env[65121]: WARNING nova.compute.manager [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received unexpected event network-vif-plugged-f119a70b-b122-4875-b1a3-5c4c7a234c29 for instance with vm_state building and task_state spawning. [ 617.649634] env[65121]: DEBUG nova.compute.manager [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 617.649737] env[65121]: DEBUG nova.compute.manager [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing instance network info cache due to event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 617.649914] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Acquiring lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.650110] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Acquired lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.650234] env[65121]: DEBUG nova.network.neutron [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 617.681972] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Releasing lock "refresh_cache-bad21c0b-6534-46b5-97a3-a26e5ecb5c96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.682261] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Received event network-vif-plugged-062568cf-fd2f-407b-8cf8-da76b66e5c30 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 617.682442] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Acquiring lock "5e915c8e-150a-4bfd-8005-a7a1a119057d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.682638] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.682785] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.682937] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] No waiting events found dispatching network-vif-plugged-062568cf-fd2f-407b-8cf8-da76b66e5c30 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 617.683106] env[65121]: WARNING nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Received unexpected event network-vif-plugged-062568cf-fd2f-407b-8cf8-da76b66e5c30 for instance with vm_state building and task_state spawning. [ 617.683257] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Received event network-changed-062568cf-fd2f-407b-8cf8-da76b66e5c30 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 617.683449] env[65121]: DEBUG nova.compute.manager [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Refreshing instance network info cache due to event network-changed-062568cf-fd2f-407b-8cf8-da76b66e5c30. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 617.683755] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Acquiring lock "refresh_cache-5e915c8e-150a-4bfd-8005-a7a1a119057d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.683755] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Acquired lock "refresh_cache-5e915c8e-150a-4bfd-8005-a7a1a119057d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.684023] env[65121]: DEBUG nova.network.neutron [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Refreshing network info cache for port 062568cf-fd2f-407b-8cf8-da76b66e5c30 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 617.863555] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5217fc3e-44ff-fe7b-3bf6-8ee1abbd2818, 'name': SearchDatastore_Task, 'duration_secs': 0.014482} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.863862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.864083] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 617.864813] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.947384] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105934, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034465} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.951612] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 617.951953] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6fc35eb-18bd-4c19-891d-741d8f5dc4e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.958876] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 617.958876] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52be2aeb-ec90-c54e-a08d-f4e7847f64c2" [ 617.958876] env[65121]: _type = "Task" [ 617.958876] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.972349] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be2aeb-ec90-c54e-a08d-f4e7847f64c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.002318] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "e2d32a5e-c350-4b2b-9243-c3b412193a82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.002955] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.022165] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb11f43-fb32-4314-9b58-4d422778ee51 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.030599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "6a4841b6-6076-4e9c-a6cf-a658957e2931" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.030897] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.042870] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a5c327-63b9-4c8a-8551-0409dcc96417 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.049160] env[65121]: DEBUG nova.network.neutron [-] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 618.081514] env[65121]: DEBUG nova.compute.manager [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 618.081721] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.083144] env[65121]: INFO nova.compute.manager [-] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Took 0.87 seconds to deallocate network for instance. [ 618.085584] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961be84e-5d62-49f0-a336-41f722caf0a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.087422] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0237e888-6c13-4d18-9b2d-c93a619924c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.102123] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 618.104533] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1c84654-987c-47ef-a8d1-961592a03130 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.107715] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b3adf0-4b71-41a4-9ebf-3dae2c4fb11b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.126878] env[65121]: DEBUG nova.compute.provider_tree [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.132716] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 618.132716] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 618.132935] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 618.133221] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 618.133221] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 618.133396] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 618.133626] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.133743] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 618.133902] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 618.135632] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 618.135632] env[65121]: DEBUG nova.virt.hardware [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 618.137367] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e453d4b-c2cf-405d-8437-4554b23bf3c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.140560] env[65121]: DEBUG oslo_vmware.api [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 618.140560] env[65121]: value = "task-5105935" [ 618.140560] env[65121]: _type = "Task" [ 618.140560] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.151202] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7f76c4-22a0-45f9-bec5-76da072daf2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.164478] env[65121]: WARNING neutronclient.v2_0.client [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.165427] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.165558] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.172930] env[65121]: DEBUG oslo_vmware.api [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105935, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.186574] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 618.192895] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 618.193592] env[65121]: WARNING neutronclient.v2_0.client [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.194320] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.194721] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.202081] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 618.203158] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ff2f9ca-909a-436d-abcb-2065b8c81159 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.231607] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 618.231607] env[65121]: value = "task-5105936" [ 618.231607] env[65121]: _type = "Task" [ 618.231607] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.243933] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105936, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.333018] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.333748] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.342925] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.343543] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.472475] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be2aeb-ec90-c54e-a08d-f4e7847f64c2, 'name': SearchDatastore_Task, 'duration_secs': 0.011232} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.472687] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.472942] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4/2700ee2c-cc57-40e0-8069-f1b7ad2df7b4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 618.473218] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.473458] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.473686] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aac34620-ee0e-4283-b1ba-273a4d306365 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.476050] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6049bae5-4259-4874-88ab-9290449288b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.483863] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 618.483863] env[65121]: value = "task-5105937" [ 618.483863] env[65121]: _type = "Task" [ 618.483863] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.488571] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.488868] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 618.490063] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d4a8168-01a0-4960-ac7b-9fdbc973c3da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.495880] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105937, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.499777] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 618.499777] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520c7ed9-cd92-5f05-8090-eb3957536674" [ 618.499777] env[65121]: _type = "Task" [ 618.499777] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.508445] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520c7ed9-cd92-5f05-8090-eb3957536674, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.557108] env[65121]: WARNING neutronclient.v2_0.client [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.557108] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.557108] env[65121]: WARNING openstack [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.586418] env[65121]: WARNING neutronclient.v2_0.client [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.586418] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.587260] env[65121]: WARNING openstack [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.599810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.642176] env[65121]: DEBUG nova.scheduler.client.report [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 618.658570] env[65121]: DEBUG oslo_vmware.api [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105935, 'name': PowerOffVM_Task, 'duration_secs': 0.294033} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.658570] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 618.658570] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 618.658570] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88d22650-97a5-4de6-8239-8e97aa34c39c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.741925] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105936, 'name': CreateVM_Task, 'duration_secs': 0.35221} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.742143] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 618.742597] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.742731] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.743082] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 618.743505] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-689a1885-7954-4492-bde7-3356df7c8f2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.752462] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 618.752462] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526ff02c-bf59-b137-6197-a0ea93431fe2" [ 618.752462] env[65121]: _type = "Task" [ 618.752462] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.762364] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526ff02c-bf59-b137-6197-a0ea93431fe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.775634] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 618.775634] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 618.775634] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Deleting the datastore file [datastore1] 0151ed85-bba4-443a-b18f-58272c973e1e {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 618.775839] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de728f8b-838e-43ba-af8f-80a37d73516a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.783020] env[65121]: DEBUG oslo_vmware.api [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for the task: (returnval){ [ 618.783020] env[65121]: value = "task-5105939" [ 618.783020] env[65121]: _type = "Task" [ 618.783020] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.792767] env[65121]: DEBUG oslo_vmware.api [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.821208] env[65121]: DEBUG nova.network.neutron [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updated VIF entry in instance network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 618.821657] env[65121]: DEBUG nova.network.neutron [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updating instance_info_cache with network_info: [{"id": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "address": "fa:16:3e:47:bb:ea", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119a70b-b1", "ovs_interfaceid": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 618.920805] env[65121]: DEBUG nova.network.neutron [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Updated VIF entry in instance network info cache for port 062568cf-fd2f-407b-8cf8-da76b66e5c30. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 618.921755] env[65121]: DEBUG nova.network.neutron [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Updating instance_info_cache with network_info: [{"id": "062568cf-fd2f-407b-8cf8-da76b66e5c30", "address": "fa:16:3e:5c:d1:31", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap062568cf-fd", "ovs_interfaceid": "062568cf-fd2f-407b-8cf8-da76b66e5c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 619.005733] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105937, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.018223] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520c7ed9-cd92-5f05-8090-eb3957536674, 'name': SearchDatastore_Task, 'duration_secs': 0.021222} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.019225] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aefa52dc-0edd-4e55-aaf6-8f9b754894c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.026638] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 619.026638] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52175061-4ac6-7b3a-85f1-98286f3b2356" [ 619.026638] env[65121]: _type = "Task" [ 619.026638] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.037808] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52175061-4ac6-7b3a-85f1-98286f3b2356, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.151912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.152505] env[65121]: DEBUG nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 619.157276] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.198s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.160337] env[65121]: INFO nova.compute.claims [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.247143] env[65121]: DEBUG nova.compute.manager [req-63b28084-9bcf-423f-8035-66b9081b0668 req-42a7ae7a-85be-414c-a306-955eb41cba62 service nova] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Received event network-vif-deleted-a059ff79-fc05-40c7-9def-3f3c476870b8 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 619.266339] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526ff02c-bf59-b137-6197-a0ea93431fe2, 'name': SearchDatastore_Task, 'duration_secs': 0.06407} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.266749] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.267317] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 619.267674] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.297966] env[65121]: DEBUG oslo_vmware.api [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Task: {'id': task-5105939, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.338874} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.298270] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 619.298820] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 619.302020] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.302020] env[65121]: INFO nova.compute.manager [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Took 1.22 seconds to destroy the instance on the hypervisor. [ 619.302020] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 619.302020] env[65121]: DEBUG nova.compute.manager [-] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 619.302020] env[65121]: DEBUG nova.network.neutron [-] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 619.302020] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.302317] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.302317] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.325168] env[65121]: DEBUG oslo_concurrency.lockutils [req-20df044d-54a4-434d-9448-dcaffa9f0789 req-b940f1f2-fd58-4d46-a1ea-b0a0a8e658e7 service nova] Releasing lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.351853] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.424149] env[65121]: DEBUG oslo_concurrency.lockutils [req-470a422e-ddb2-4487-b0f7-baf5eb4e9108 req-f4a43056-b750-4889-9c0f-8aa8763500f8 service nova] Releasing lock "refresh_cache-5e915c8e-150a-4bfd-8005-a7a1a119057d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.499261] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105937, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546877} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.499804] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4/2700ee2c-cc57-40e0-8069-f1b7ad2df7b4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 619.500465] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 619.500966] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e148de4f-8695-4a73-b288-71affebfb051 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.510549] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 619.510549] env[65121]: value = "task-5105940" [ 619.510549] env[65121]: _type = "Task" [ 619.510549] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.521285] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.540871] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52175061-4ac6-7b3a-85f1-98286f3b2356, 'name': SearchDatastore_Task, 'duration_secs': 0.024296} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.541159] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.541499] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] aa9b6708-c53c-4117-9b75-9d506f393395/aa9b6708-c53c-4117-9b75-9d506f393395.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 619.541809] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.542032] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 619.542272] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abb95f41-33fc-4a7d-ac56-c01d6cc6ef3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.546020] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-11bfb2b0-85f1-428c-8e08-c36174363b1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.555876] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 619.555876] env[65121]: value = "task-5105941" [ 619.555876] env[65121]: _type = "Task" [ 619.555876] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.562418] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 619.562693] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 619.563928] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b64bcb4-f44f-4a77-835c-588f6a93e806 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.570099] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105941, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.574331] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 619.574331] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ce6108-5ef0-5196-e750-85f61228dcce" [ 619.574331] env[65121]: _type = "Task" [ 619.574331] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.585433] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ce6108-5ef0-5196-e750-85f61228dcce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.665703] env[65121]: DEBUG nova.compute.utils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 619.667134] env[65121]: DEBUG nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 619.670437] env[65121]: DEBUG nova.network.neutron [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 619.670796] env[65121]: WARNING neutronclient.v2_0.client [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.671164] env[65121]: WARNING neutronclient.v2_0.client [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.671762] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.672112] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.024843] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100933} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.025163] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 620.029119] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3bde56-d82f-4534-9e56-3b9b3f765b7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.058629] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Reconfiguring VM instance instance-00000006 to attach disk [datastore2] 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4/2700ee2c-cc57-40e0-8069-f1b7ad2df7b4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 620.060579] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1ed9585-2ddf-4fed-806c-f4dc502ef92d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.077141] env[65121]: DEBUG nova.network.neutron [-] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 620.096049] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105941, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533161} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.100523] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] aa9b6708-c53c-4117-9b75-9d506f393395/aa9b6708-c53c-4117-9b75-9d506f393395.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 620.100756] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 620.101507] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 620.101507] env[65121]: value = "task-5105942" [ 620.101507] env[65121]: _type = "Task" [ 620.101507] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.101696] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ce6108-5ef0-5196-e750-85f61228dcce, 'name': SearchDatastore_Task, 'duration_secs': 0.011412} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.101835] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e75ec893-a5a0-41af-baee-2cfcbe87fde1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.108086] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6747cdd6-9343-4100-8d35-c82036218f5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.120428] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.123692] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 620.123692] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52239039-94a5-a91a-959f-dc56d194e20e" [ 620.123692] env[65121]: _type = "Task" [ 620.123692] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.123692] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 620.123692] env[65121]: value = "task-5105943" [ 620.123692] env[65121]: _type = "Task" [ 620.123692] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.138245] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52239039-94a5-a91a-959f-dc56d194e20e, 'name': SearchDatastore_Task, 'duration_secs': 0.01135} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.142934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.143438] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] bad21c0b-6534-46b5-97a3-a26e5ecb5c96/bad21c0b-6534-46b5-97a3-a26e5ecb5c96.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 620.143867] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105943, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.144205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.144474] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 620.145081] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efd5b994-7498-4b34-be5d-15002af8a4a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.147830] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04932e0e-1d21-4f44-81ce-79f636474815 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.158111] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 620.158111] env[65121]: value = "task-5105944" [ 620.158111] env[65121]: _type = "Task" [ 620.158111] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.160039] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 620.160304] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 620.166134] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83d36f7c-31a0-4a26-bdcb-b15260ebfaf1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.176410] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.177642] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 620.177642] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a85801-86b2-5157-1fbd-ab522bbf22bd" [ 620.177642] env[65121]: _type = "Task" [ 620.177642] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.187885] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a85801-86b2-5157-1fbd-ab522bbf22bd, 'name': SearchDatastore_Task, 'duration_secs': 0.010645} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.188063] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85ba6762-b0e5-4e40-be1b-c90f85f02380 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.191438] env[65121]: DEBUG nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 620.198576] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 620.198576] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52203319-d533-a06f-7b60-80af81dd818f" [ 620.198576] env[65121]: _type = "Task" [ 620.198576] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.209572] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52203319-d533-a06f-7b60-80af81dd818f, 'name': SearchDatastore_Task, 'duration_secs': 0.009699} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.209963] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.210272] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 5e915c8e-150a-4bfd-8005-a7a1a119057d/5e915c8e-150a-4bfd-8005-a7a1a119057d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 620.210917] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.210917] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 620.211067] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-199da029-d5a3-413a-965f-20261806bfef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.213450] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f52ce40-acc4-4d0c-982d-e2ac7f75e35c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.223652] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 620.223652] env[65121]: value = "task-5105945" [ 620.223652] env[65121]: _type = "Task" [ 620.223652] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.226450] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 620.226450] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 620.229617] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b68c366a-262b-4fab-8d9b-a8ce57b643f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.242021] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105945, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.242021] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 620.242021] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527fda47-0602-9dd0-0566-d646137f06f1" [ 620.242021] env[65121]: _type = "Task" [ 620.242021] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.253954] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527fda47-0602-9dd0-0566-d646137f06f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.299304] env[65121]: DEBUG nova.policy [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 620.571918] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4728bd15-027e-4a35-abc4-571082694433 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.583567] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb231fea-6dab-4449-acd6-cbb69ca1ec17 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.589623] env[65121]: INFO nova.compute.manager [-] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Took 1.29 seconds to deallocate network for instance. [ 620.641617] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0927eb27-7306-43ef-88e1-dfabb78b7e4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.658628] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61ee863-1f40-41ad-a7bc-f8e7788e89d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.663617] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105943, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075376} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.663818] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105942, 'name': ReconfigVM_Task, 'duration_secs': 0.558544} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.664846] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 620.665171] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Reconfigured VM instance instance-00000006 to attach disk [datastore2] 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4/2700ee2c-cc57-40e0-8069-f1b7ad2df7b4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 620.669848] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331fbeca-fc92-473f-9576-84237285afbc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.673221] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3fff2b52-5f53-434b-a100-e55a6985170f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.687337] env[65121]: DEBUG nova.compute.provider_tree [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.697305] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523598} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.711904] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] bad21c0b-6534-46b5-97a3-a26e5ecb5c96/bad21c0b-6534-46b5-97a3-a26e5ecb5c96.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 620.712191] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 620.727420] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] aa9b6708-c53c-4117-9b75-9d506f393395/aa9b6708-c53c-4117-9b75-9d506f393395.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 620.728903] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9089be44-8cfc-4708-bfa6-3bc55552780d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.732687] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea1520c9-6af9-41e4-a604-6719b71c81e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.756205] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 620.756205] env[65121]: value = "task-5105946" [ 620.756205] env[65121]: _type = "Task" [ 620.756205] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.773629] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 620.773629] env[65121]: value = "task-5105947" [ 620.773629] env[65121]: _type = "Task" [ 620.773629] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.774888] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105945, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.777301] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 620.777301] env[65121]: value = "task-5105948" [ 620.777301] env[65121]: _type = "Task" [ 620.777301] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.793825] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527fda47-0602-9dd0-0566-d646137f06f1, 'name': SearchDatastore_Task, 'duration_secs': 0.019981} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.797371] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad7a3907-0c28-4489-aaa7-64e695dbbb4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.814690] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105946, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.814882] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105947, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.815117] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105948, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.822458] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 620.822458] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52be793d-1f91-5259-06c4-af3a80a5ae95" [ 620.822458] env[65121]: _type = "Task" [ 620.822458] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.835090] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be793d-1f91-5259-06c4-af3a80a5ae95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.134357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.193920] env[65121]: DEBUG nova.scheduler.client.report [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 621.230616] env[65121]: DEBUG nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 621.243266] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105945, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.79948} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.243613] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 5e915c8e-150a-4bfd-8005-a7a1a119057d/5e915c8e-150a-4bfd-8005-a7a1a119057d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 621.243821] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 621.244112] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4da9c748-b6e8-4ed5-b24b-92c0c5d4e3a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.254031] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 621.254031] env[65121]: value = "task-5105949" [ 621.254031] env[65121]: _type = "Task" [ 621.254031] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.259826] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 621.260076] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 621.260226] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 621.260397] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 621.260600] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 621.260675] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 621.260923] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.261079] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 621.261238] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 621.261428] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 621.261582] env[65121]: DEBUG nova.virt.hardware [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 621.262501] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3c4de2-d38b-47be-a338-d68057d7998a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.272481] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.280620] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed229632-b19a-402a-8e30-f3c878df580d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.285974] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105946, 'name': Rename_Task, 'duration_secs': 0.236809} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.289489] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 621.293277] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64b844e2-e1e9-4050-a018-425d0e98f596 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.312638] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105947, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118225} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.316066] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.316428] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105948, 'name': ReconfigVM_Task, 'duration_secs': 0.377214} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.318313] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82281692-1225-4991-ba9b-fcbda224d958 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.320983] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Reconfigured VM instance instance-00000007 to attach disk [datastore2] aa9b6708-c53c-4117-9b75-9d506f393395/aa9b6708-c53c-4117-9b75-9d506f393395.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 621.321695] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 621.321695] env[65121]: value = "task-5105950" [ 621.321695] env[65121]: _type = "Task" [ 621.321695] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.321936] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-530e4c91-52cf-4228-a667-d8ffe797a71f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.347635] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] bad21c0b-6534-46b5-97a3-a26e5ecb5c96/bad21c0b-6534-46b5-97a3-a26e5ecb5c96.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.349714] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31951237-baf5-4462-873e-838f0df9093b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.364198] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 621.364198] env[65121]: value = "task-5105951" [ 621.364198] env[65121]: _type = "Task" [ 621.364198] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.369077] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105950, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.376317] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be793d-1f91-5259-06c4-af3a80a5ae95, 'name': SearchDatastore_Task, 'duration_secs': 0.056021} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.378292] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 621.378539] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 621.378864] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 621.378864] env[65121]: value = "task-5105952" [ 621.378864] env[65121]: _type = "Task" [ 621.378864] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.379077] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a339be57-6d88-4f84-bba0-458f6e470138 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.387416] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105951, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.396907] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105952, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.398751] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 621.398751] env[65121]: value = "task-5105953" [ 621.398751] env[65121]: _type = "Task" [ 621.398751] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.409308] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105953, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.473193] env[65121]: DEBUG nova.network.neutron [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Successfully created port: 13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 621.698419] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 621.699186] env[65121]: DEBUG nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 621.703576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.448s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 621.705118] env[65121]: INFO nova.compute.claims [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.767787] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063246} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.768295] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.770226] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c93869-607e-4f12-882a-d943bcf0e0cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.806607] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] 5e915c8e-150a-4bfd-8005-a7a1a119057d/5e915c8e-150a-4bfd-8005-a7a1a119057d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.807149] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32d6ffd9-c39d-4f3f-ab6f-a29728e4ba28 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.829213] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 621.829213] env[65121]: value = "task-5105954" [ 621.829213] env[65121]: _type = "Task" [ 621.829213] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.835927] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105950, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.844597] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105954, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.879811] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105951, 'name': Rename_Task, 'duration_secs': 0.170142} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.880194] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 621.880599] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25977175-9529-4577-ae2a-25fde2fe9acf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.894637] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105952, 'name': ReconfigVM_Task, 'duration_secs': 0.41796} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.895995] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Reconfigured VM instance instance-00000008 to attach disk [datastore2] bad21c0b-6534-46b5-97a3-a26e5ecb5c96/bad21c0b-6534-46b5-97a3-a26e5ecb5c96.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 621.896851] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 621.896851] env[65121]: value = "task-5105955" [ 621.896851] env[65121]: _type = "Task" [ 621.896851] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.897313] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc5790d1-849f-46b0-88f9-7229cf17829c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.912919] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.919703] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 621.919703] env[65121]: value = "task-5105956" [ 621.919703] env[65121]: _type = "Task" [ 621.919703] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.919703] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105953, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.931915] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105956, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.209506] env[65121]: DEBUG nova.compute.utils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 622.215442] env[65121]: DEBUG nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 622.215442] env[65121]: DEBUG nova.network.neutron [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 622.215442] env[65121]: WARNING neutronclient.v2_0.client [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.215442] env[65121]: WARNING neutronclient.v2_0.client [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.216610] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.216610] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.337786] env[65121]: DEBUG oslo_vmware.api [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105950, 'name': PowerOnVM_Task, 'duration_secs': 0.718642} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.338548] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 622.338824] env[65121]: INFO nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Took 16.59 seconds to spawn the instance on the hypervisor. [ 622.339095] env[65121]: DEBUG nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 622.339938] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33f0bd1-d8fc-4792-acf4-b8c7beeecafa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.346170] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105954, 'name': ReconfigVM_Task, 'duration_secs': 0.44101} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.346854] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Reconfigured VM instance instance-00000009 to attach disk [datastore2] 5e915c8e-150a-4bfd-8005-a7a1a119057d/5e915c8e-150a-4bfd-8005-a7a1a119057d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.347596] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee923ee7-dd2f-4b41-816e-dbe6033e1ae3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.359740] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 622.359740] env[65121]: value = "task-5105957" [ 622.359740] env[65121]: _type = "Task" [ 622.359740] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.371640] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105957, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.418132] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105953, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549528} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.418423] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105955, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.418710] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 622.418954] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 622.419249] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-430cbd55-350e-4221-aff6-266d60451d07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.432116] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105956, 'name': Rename_Task, 'duration_secs': 0.290299} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.433992] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 622.434384] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 622.434384] env[65121]: value = "task-5105958" [ 622.434384] env[65121]: _type = "Task" [ 622.434384] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.434870] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04d6889b-1b83-44fb-84a7-f96e45d4a77f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.447713] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105958, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.449794] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 622.449794] env[65121]: value = "task-5105959" [ 622.449794] env[65121]: _type = "Task" [ 622.449794] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.461494] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105959, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.719833] env[65121]: DEBUG nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 622.864017] env[65121]: INFO nova.compute.manager [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Took 29.26 seconds to build instance. [ 622.878872] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105957, 'name': Rename_Task, 'duration_secs': 0.256642} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.881666] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 622.882258] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67532f39-8998-4a82-a0df-b38f3e8e883b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.890464] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 622.890464] env[65121]: value = "task-5105960" [ 622.890464] env[65121]: _type = "Task" [ 622.890464] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.910574] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.923190] env[65121]: DEBUG oslo_vmware.api [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5105955, 'name': PowerOnVM_Task, 'duration_secs': 0.574144} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.923668] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 622.923811] env[65121]: INFO nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Took 15.18 seconds to spawn the instance on the hypervisor. [ 622.924541] env[65121]: DEBUG nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 622.925728] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ae9ed1-5650-4d71-a1f2-0cfe5b9cdeb8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.950427] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105958, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071724} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.956881] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 622.958230] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca361451-0fbe-46c1-802e-ae8c75631bb5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.987385] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 622.987795] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105959, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.991731] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7a23880-fb51-46ad-bfc2-95ced5fadef4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.016238] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 623.016238] env[65121]: value = "task-5105961" [ 623.016238] env[65121]: _type = "Task" [ 623.016238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.032283] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105961, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.109473] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae9ea72-7da4-4e36-903d-98940535381d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.118354] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df257fe-5c38-462a-8071-ac2375505ce3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.150692] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c9716f-f59b-430a-9966-bb3ab82a929f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.159260] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82318a23-bcb8-455f-9255-cb7d47b63f4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.176206] env[65121]: DEBUG nova.compute.provider_tree [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.371622] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7bb83e88-0b60-4115-a1b5-d3ddfe2ae0cb tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.777s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 623.408879] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105960, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.463713] env[65121]: INFO nova.compute.manager [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Took 29.80 seconds to build instance. [ 623.475281] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105959, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.533297] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.557890] env[65121]: DEBUG nova.network.neutron [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Successfully updated port: 13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 623.623820] env[65121]: DEBUG nova.policy [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c468565c58704d9084cfebdaa2897492', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce90017874734825b9fbdce06eb12a00', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 623.677941] env[65121]: DEBUG nova.scheduler.client.report [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 623.731075] env[65121]: DEBUG nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 623.766096] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 623.766443] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 623.766647] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 623.766895] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 623.767094] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 623.767289] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 623.767559] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.767763] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 623.767978] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 623.768248] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 623.768486] env[65121]: DEBUG nova.virt.hardware [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 623.771033] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567dd7be-0d06-4b69-a0a3-fe9ed0aa80ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.782189] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee01421-de64-4a54-9d61-14d39bb71ec0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.876200] env[65121]: DEBUG nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 623.909451] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105960, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.966215] env[65121]: DEBUG oslo_concurrency.lockutils [None req-04acffe3-c413-45d4-b34a-a762f1997e14 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "aa9b6708-c53c-4117-9b75-9d506f393395" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.307s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 623.971563] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105959, 'name': PowerOnVM_Task, 'duration_secs': 1.150656} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.972636] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 623.972636] env[65121]: INFO nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Took 13.23 seconds to spawn the instance on the hypervisor. [ 623.972636] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 623.973450] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7137d7-3b1b-495a-9c2b-b0aef5bf78e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.030228] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105961, 'name': ReconfigVM_Task, 'duration_secs': 0.613131} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.030228] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 624.030974] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b5d19df-dc8b-4f90-87e6-394f2943fcb1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.040466] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 624.040466] env[65121]: value = "task-5105962" [ 624.040466] env[65121]: _type = "Task" [ 624.040466] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.051174] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105962, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.065936] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.066220] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.066632] env[65121]: DEBUG nova.network.neutron [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 624.185840] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.187253] env[65121]: DEBUG nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 624.191190] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.678s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.193095] env[65121]: INFO nova.compute.claims [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.406951] env[65121]: DEBUG oslo_vmware.api [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5105960, 'name': PowerOnVM_Task, 'duration_secs': 1.085992} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.408258] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.408534] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 624.408724] env[65121]: INFO nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Took 11.24 seconds to spawn the instance on the hypervisor. [ 624.408885] env[65121]: DEBUG nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 624.409721] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9567a717-91bb-42a0-86c3-0117500220dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.431332] env[65121]: DEBUG nova.network.neutron [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Successfully created port: 3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 624.481486] env[65121]: DEBUG nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 624.495903] env[65121]: INFO nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Took 30.22 seconds to build instance. [ 624.556131] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105962, 'name': Rename_Task, 'duration_secs': 0.321697} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.558301] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 624.559293] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8db4c601-74aa-4cc6-9e20-399e91ad0c58 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.568826] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 624.568826] env[65121]: value = "task-5105963" [ 624.568826] env[65121]: _type = "Task" [ 624.568826] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.569301] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.569921] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 624.590530] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105963, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.692944] env[65121]: DEBUG nova.compute.utils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 624.695732] env[65121]: DEBUG nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 624.797285] env[65121]: DEBUG nova.network.neutron [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 624.936618] env[65121]: INFO nova.compute.manager [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Took 30.61 seconds to build instance. [ 624.998889] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.738s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.011072] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.092939] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105963, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.200025] env[65121]: DEBUG nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 625.439080] env[65121]: DEBUG oslo_concurrency.lockutils [None req-61b2abcd-5aa3-4979-ba4c-b82fe2b4982d tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.116s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.504292] env[65121]: DEBUG nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 625.526444] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efa4f94-a4f8-4983-b23f-d9f0ff0d106e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.538436] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7fb770-7193-42ae-b398-580f2623234d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.575500] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a27f57e-93b3-444f-a084-13c7207b79c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.594289] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174bfee3-13b2-4afb-ad73-9b5721f7e460 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.598876] env[65121]: DEBUG oslo_vmware.api [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5105963, 'name': PowerOnVM_Task, 'duration_secs': 0.577529} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.599603] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 625.599958] env[65121]: DEBUG nova.compute.manager [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 625.601358] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863d861a-2bfc-4fb3-9d88-67abe1996956 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.612390] env[65121]: DEBUG nova.compute.provider_tree [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.684391] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.684826] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.873871] env[65121]: DEBUG nova.compute.manager [req-298a7d53-d8be-433c-ac17-a42aacdf6566 req-55c37346-2487-48da-b6bf-9818e1931756 service nova] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Received event network-vif-deleted-cf0d4936-618c-45a3-af9c-3f745b2cf316 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 626.033845] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.117097] env[65121]: DEBUG nova.scheduler.client.report [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 626.131857] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.211143] env[65121]: DEBUG nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 626.244247] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 626.244544] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 626.245303] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 626.245303] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 626.245303] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 626.245303] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 626.245506] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.245542] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 626.245729] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 626.245845] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 626.246012] env[65121]: DEBUG nova.virt.hardware [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 626.247502] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830230e0-517f-4299-b6d2-25af8ecab200 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.256588] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6451d76-b63b-4410-a088-dd15eb94f3d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.271506] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 626.277848] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Creating folder: Project (f18aa919bbcf4562be99ba9defcd5256). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.278283] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc1e6bee-3f02-404b-8911-622647b2381d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.292419] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Created folder: Project (f18aa919bbcf4562be99ba9defcd5256) in parent group-v993268. [ 626.292419] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Creating folder: Instances. Parent ref: group-v993295. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.292419] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba421c24-549f-40e3-9f4b-b9289e1b3f1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.294501] env[65121]: DEBUG nova.network.neutron [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Successfully updated port: 3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 626.308574] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Created folder: Instances in parent group-v993295. [ 626.308951] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 626.309329] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 626.309551] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16f628a4-89cf-44d0-8f7b-7a0fadfe8f6c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.331977] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 626.331977] env[65121]: value = "task-5105966" [ 626.331977] env[65121]: _type = "Task" [ 626.331977] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.341805] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105966, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.496495] env[65121]: WARNING neutronclient.v2_0.client [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.497506] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.497951] env[65121]: WARNING openstack [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.625125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 626.625125] env[65121]: DEBUG nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 626.626679] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.762s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 626.632462] env[65121]: INFO nova.compute.claims [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.797339] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.797523] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 626.797741] env[65121]: DEBUG nova.network.neutron [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 626.844853] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105966, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.045390] env[65121]: DEBUG nova.network.neutron [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 627.139118] env[65121]: DEBUG nova.compute.utils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 627.143405] env[65121]: DEBUG nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 627.143578] env[65121]: DEBUG nova.network.neutron [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 627.144152] env[65121]: WARNING neutronclient.v2_0.client [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.144226] env[65121]: WARNING neutronclient.v2_0.client [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.144816] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.145429] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.225020] env[65121]: DEBUG nova.policy [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af00bd582c0843949491bbcecbfcd2dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72237a0d762645588c41231b0a34a796', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 627.303739] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.304292] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.345683] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105966, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.491116] env[65121]: DEBUG nova.network.neutron [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 627.550586] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 627.550586] env[65121]: DEBUG nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Instance network_info: |[{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 627.551028] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:ed:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9ec24851-7bb6-426b-b28f-f7b246df1713', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13f8bd13-f8f0-4818-8dd6-845e9644a395', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 627.557870] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Creating folder: Project (35312a302644426f98f127e89a067e75). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 627.558181] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae6352b3-33b1-4e34-ae32-fe6550b34f71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.572144] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Created folder: Project (35312a302644426f98f127e89a067e75) in parent group-v993268. [ 627.572355] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Creating folder: Instances. Parent ref: group-v993298. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 627.573822] env[65121]: DEBUG nova.network.neutron [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Successfully created port: 4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 627.578624] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-495bf832-97af-4252-b7aa-b5f5f98367aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.589936] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.591156] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.599682] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Created folder: Instances in parent group-v993298. [ 627.599943] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 627.600192] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 627.600558] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e352248b-76d5-4b8e-bab8-6b0352ed6a31 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.630405] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 627.630405] env[65121]: value = "task-5105969" [ 627.630405] env[65121]: _type = "Task" [ 627.630405] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.651756] env[65121]: DEBUG nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 627.656284] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105969, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.706929] env[65121]: WARNING neutronclient.v2_0.client [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.707815] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.708217] env[65121]: WARNING openstack [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.850204] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105966, 'name': CreateVM_Task, 'duration_secs': 1.059147} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.850204] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 627.850204] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.850204] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.850527] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 627.850872] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-687db5f3-42cb-46a2-b58b-87b8210a98c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.859589] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 627.859589] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529959ff-d772-9edd-1455-e339b791786f" [ 627.859589] env[65121]: _type = "Task" [ 627.859589] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.870425] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529959ff-d772-9edd-1455-e339b791786f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.049954] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc319a6-df93-4322-acff-eebf44b02533 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.061658] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cafdc51-0e34-4cbc-bf6a-a95856e28d23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.113192] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0c1905-caeb-41d4-b976-57b334c26795 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.124941] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152d5436-fa90-4441-8248-7709e61cf7c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.148468] env[65121]: DEBUG nova.compute.provider_tree [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.153823] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105969, 'name': CreateVM_Task, 'duration_secs': 0.48968} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.154383] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 628.155244] env[65121]: WARNING neutronclient.v2_0.client [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.156017] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.376513] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529959ff-d772-9edd-1455-e339b791786f, 'name': SearchDatastore_Task, 'duration_secs': 0.01319} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.376980] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.377371] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 628.377736] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.377969] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.378273] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 628.378682] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.379188] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 628.379450] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c096ea46-22b8-42b0-a361-1367c17e0e75 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.382601] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f219834-9d27-4d99-b0e2-a53d55cb211c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.389875] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 628.389875] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523041cd-6e01-7d26-7a84-43011355552c" [ 628.389875] env[65121]: _type = "Task" [ 628.389875] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.401534] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 628.401851] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 628.403892] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6df9a71c-6307-48a1-bd5d-f32005a46950 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.413487] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523041cd-6e01-7d26-7a84-43011355552c, 'name': SearchDatastore_Task, 'duration_secs': 0.01278} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.414563] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.414870] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 628.415209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.418258] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 628.418258] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52646f37-81a9-2ac0-fa9e-77f11b89e7aa" [ 628.418258] env[65121]: _type = "Task" [ 628.418258] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.428301] env[65121]: DEBUG nova.network.neutron [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updating instance_info_cache with network_info: [{"id": "3a66c1d6-9549-459d-bb1b-66781725a57f", "address": "fa:16:3e:be:bf:37", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66c1d6-95", "ovs_interfaceid": "3a66c1d6-9549-459d-bb1b-66781725a57f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 628.435423] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52646f37-81a9-2ac0-fa9e-77f11b89e7aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.655847] env[65121]: DEBUG nova.scheduler.client.report [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 628.664670] env[65121]: DEBUG nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 628.699548] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 628.699863] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 628.701456] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 628.701456] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 628.701456] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 628.703496] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 628.703496] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.703496] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 628.703496] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 628.703496] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 628.705769] env[65121]: DEBUG nova.virt.hardware [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 628.705769] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd17a03-58ed-4958-8c14-5727ac403f47 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.714111] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f30e77-7a43-4688-b2fd-a37df90d80b0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.933225] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52646f37-81a9-2ac0-fa9e-77f11b89e7aa, 'name': SearchDatastore_Task, 'duration_secs': 0.011018} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.934100] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-237116af-bda6-4315-bfb7-95ae8dab49d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.936858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Releasing lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.937243] env[65121]: DEBUG nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Instance network_info: |[{"id": "3a66c1d6-9549-459d-bb1b-66781725a57f", "address": "fa:16:3e:be:bf:37", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66c1d6-95", "ovs_interfaceid": "3a66c1d6-9549-459d-bb1b-66781725a57f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 628.937766] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:bf:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a66c1d6-9549-459d-bb1b-66781725a57f', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 628.946943] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Creating folder: Project (ce90017874734825b9fbdce06eb12a00). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 628.948157] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-49e67158-483f-467f-8f1a-dffe8d6fc711 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.951653] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 628.951653] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52222a11-c690-3b3a-70b5-b3705d22cd59" [ 628.951653] env[65121]: _type = "Task" [ 628.951653] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.962075] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52222a11-c690-3b3a-70b5-b3705d22cd59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.964507] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Created folder: Project (ce90017874734825b9fbdce06eb12a00) in parent group-v993268. [ 628.965170] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Creating folder: Instances. Parent ref: group-v993301. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 628.965170] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f96aea55-f845-4306-95c9-dd74ff5e7148 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.976229] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Created folder: Instances in parent group-v993301. [ 628.976534] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 628.976833] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 628.977087] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e780a1e5-280c-4628-9e91-b5c9eae64c76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.999043] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 628.999043] env[65121]: value = "task-5105972" [ 628.999043] env[65121]: _type = "Task" [ 628.999043] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.008945] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105972, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.162131] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.162665] env[65121]: DEBUG nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 629.165839] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.758s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 629.166796] env[65121]: INFO nova.compute.claims [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.262749] env[65121]: DEBUG nova.network.neutron [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Successfully updated port: 4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 629.468083] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52222a11-c690-3b3a-70b5-b3705d22cd59, 'name': SearchDatastore_Task, 'duration_secs': 0.012688} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.468083] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.468083] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 6887bb66-7186-4d58-be59-d0c53c3221c6/6887bb66-7186-4d58-be59-d0c53c3221c6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 629.468083] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.468438] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 629.468702] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5903e6d5-24e8-4023-ab7d-039355456a04 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.471495] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a74d0cfd-6de5-4cb1-8611-0b4d705260dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.480192] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 629.480192] env[65121]: value = "task-5105973" [ 629.480192] env[65121]: _type = "Task" [ 629.480192] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.487034] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 629.487527] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 629.489282] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1719f0b-6ec7-429a-8412-f210152127cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.497069] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.505938] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 629.505938] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ea41f1-b452-452b-9c27-4373ce88ae4b" [ 629.505938] env[65121]: _type = "Task" [ 629.505938] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.515890] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105972, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.524025] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ea41f1-b452-452b-9c27-4373ce88ae4b, 'name': SearchDatastore_Task, 'duration_secs': 0.010784} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.524025] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6e3e541-5bbc-42f0-91a8-c684dea853ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.532874] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 629.532874] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fffaa3-6882-1ecf-cd01-3a70930ba8ce" [ 629.532874] env[65121]: _type = "Task" [ 629.532874] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.548599] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fffaa3-6882-1ecf-cd01-3a70930ba8ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.673033] env[65121]: DEBUG nova.compute.utils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 629.677034] env[65121]: DEBUG nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 629.677034] env[65121]: DEBUG nova.network.neutron [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 629.677034] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.677034] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.678045] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.678521] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.765080] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.765278] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.765452] env[65121]: DEBUG nova.network.neutron [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 629.994716] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105973, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.014135] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105972, 'name': CreateVM_Task, 'duration_secs': 0.627476} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.014349] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 630.014868] env[65121]: WARNING neutronclient.v2_0.client [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.015611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.015611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.015954] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 630.016128] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d346051-d075-4545-8aa6-e2326b2106fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.024362] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 630.024362] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528bcce6-07e0-1904-f955-ece6fc1c6f45" [ 630.024362] env[65121]: _type = "Task" [ 630.024362] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.035463] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528bcce6-07e0-1904-f955-ece6fc1c6f45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.045197] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fffaa3-6882-1ecf-cd01-3a70930ba8ce, 'name': SearchDatastore_Task, 'duration_secs': 0.027101} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.045502] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.045776] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] cc002961-d742-4255-88c7-f0a5a39424b1/cc002961-d742-4255-88c7-f0a5a39424b1.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 630.046058] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f77f2db-18d6-459c-9312-171fb4485261 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.055127] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 630.055127] env[65121]: value = "task-5105974" [ 630.055127] env[65121]: _type = "Task" [ 630.055127] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.065793] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105974, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.176822] env[65121]: DEBUG nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 630.246289] env[65121]: DEBUG nova.policy [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a22fb4b065945e2a779678b83dfdc5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'def20b65ac8341c1a60a747260589d5a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 630.269696] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.270438] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.498754] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553092} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.499417] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 6887bb66-7186-4d58-be59-d0c53c3221c6/6887bb66-7186-4d58-be59-d0c53c3221c6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 630.499417] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 630.499617] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b37c3c3e-6a24-48ab-a4e7-e3faec86aafd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.508853] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 630.508853] env[65121]: value = "task-5105975" [ 630.508853] env[65121]: _type = "Task" [ 630.508853] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.521073] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105975, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.537294] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528bcce6-07e0-1904-f955-ece6fc1c6f45, 'name': SearchDatastore_Task, 'duration_secs': 0.013484} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.540763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.541208] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 630.541510] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.541673] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.541849] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 630.542395] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3efde5cb-21e6-455f-891f-a84168c3ac6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.554460] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 630.554460] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 630.555071] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59310586-7119-4065-904f-1d1df888d401 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.557822] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf03ce37-2d4e-4677-8a5e-a1d13a3d0923 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.568432] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 630.568432] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529546c6-e098-1fa4-52b0-04b1a556a060" [ 630.568432] env[65121]: _type = "Task" [ 630.568432] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.574964] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105974, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48329} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.579209] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] cc002961-d742-4255-88c7-f0a5a39424b1/cc002961-d742-4255-88c7-f0a5a39424b1.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 630.579438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 630.580361] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3cf0a8-aef8-4795-a629-1300264f250f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.584291] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3248020a-da77-467d-a023-4c52d634661e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.594537] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529546c6-e098-1fa4-52b0-04b1a556a060, 'name': SearchDatastore_Task, 'duration_secs': 0.01082} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.622802] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 630.622802] env[65121]: value = "task-5105976" [ 630.622802] env[65121]: _type = "Task" [ 630.622802] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.623044] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21276d19-c428-4983-9810-4df81af56e76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.626011] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fec27b7-e332-4c78-9b8d-ffefdc88e6b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.634894] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 630.634894] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521914d7-baed-7bcd-ba3e-8264fce26b70" [ 630.634894] env[65121]: _type = "Task" [ 630.634894] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.640295] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105976, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.644415] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52a5349-13c9-4a95-a1fe-d10c0ba01601 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.656150] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521914d7-baed-7bcd-ba3e-8264fce26b70, 'name': SearchDatastore_Task, 'duration_secs': 0.010578} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.664843] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.664843] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] f1a5fdce-a940-43d4-83d7-716786c9ac34/f1a5fdce-a940-43d4-83d7-716786c9ac34.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 630.666347] env[65121]: DEBUG nova.compute.provider_tree [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.667931] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8efc4b29-155b-44f9-88a7-ac62abb853d3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.678990] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 630.678990] env[65121]: value = "task-5105977" [ 630.678990] env[65121]: _type = "Task" [ 630.678990] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.692534] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105977, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.023219] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105975, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09009} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.023665] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 631.025116] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fbd2f10-22bd-4dba-bcd4-b9a6dcb0eec4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.056923] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 6887bb66-7186-4d58-be59-d0c53c3221c6/6887bb66-7186-4d58-be59-d0c53c3221c6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 631.057768] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da672b08-7455-4b6d-8a4c-8f0031afd9a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.088272] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 631.088272] env[65121]: value = "task-5105978" [ 631.088272] env[65121]: _type = "Task" [ 631.088272] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.100590] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105978, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.144074] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105976, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09686} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.144388] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 631.146134] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0629934a-4e41-4ea6-adcf-25b22a6a93da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.172731] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] cc002961-d742-4255-88c7-f0a5a39424b1/cc002961-d742-4255-88c7-f0a5a39424b1.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 631.174880] env[65121]: DEBUG nova.scheduler.client.report [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 631.177912] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34f3ca25-7c51-4fe3-9470-0deca59674fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.204440] env[65121]: DEBUG nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 631.220499] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 631.220499] env[65121]: value = "task-5105979" [ 631.220499] env[65121]: _type = "Task" [ 631.220499] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.220692] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105977, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488802} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.221511] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] f1a5fdce-a940-43d4-83d7-716786c9ac34/f1a5fdce-a940-43d4-83d7-716786c9ac34.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 631.221511] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 631.224500] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df0f6ffa-93ac-4c07-9e83-a29b109ec6cc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.233745] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 631.233745] env[65121]: value = "task-5105980" [ 631.233745] env[65121]: _type = "Task" [ 631.233745] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.237074] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105979, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.239666] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 631.239930] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 631.240113] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 631.240276] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 631.240405] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 631.240537] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 631.240732] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.240865] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 631.241026] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 631.241175] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 631.241373] env[65121]: DEBUG nova.virt.hardware [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 631.245589] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e08a5e6-2319-4f41-b793-fecbdf06d9b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.255281] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105980, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.260039] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a4e069-b271-4852-9b11-c5e446667b89 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.343216] env[65121]: DEBUG nova.network.neutron [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 631.591767] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.591767] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.616110] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105978, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.705152] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.707467] env[65121]: DEBUG nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 631.709329] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 14.172s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 631.734760] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105979, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.750308] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105980, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112499} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.750308] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 631.751051] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40584374-9fa0-402d-9781-f44d23f1b213 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.791476] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] f1a5fdce-a940-43d4-83d7-716786c9ac34/f1a5fdce-a940-43d4-83d7-716786c9ac34.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 631.792150] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-355d5682-a7ef-445e-8013-2b1278b7c139 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.818500] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 631.818500] env[65121]: value = "task-5105981" [ 631.818500] env[65121]: _type = "Task" [ 631.818500] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.818946] env[65121]: DEBUG nova.network.neutron [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Successfully created port: ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 631.832994] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105981, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.967730] env[65121]: WARNING neutronclient.v2_0.client [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.968323] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.968665] env[65121]: WARNING openstack [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.102890] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105978, 'name': ReconfigVM_Task, 'duration_secs': 0.810351} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.102890] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 6887bb66-7186-4d58-be59-d0c53c3221c6/6887bb66-7186-4d58-be59-d0c53c3221c6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 632.104559] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04dd51d7-0c94-41f8-9c1d-6364a2f308b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.112886] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 632.112886] env[65121]: value = "task-5105982" [ 632.112886] env[65121]: _type = "Task" [ 632.112886] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.123955] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105982, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.205591] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.205591] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.206835] env[65121]: DEBUG nova.compute.manager [req-687ab517-d6d2-43f5-8781-8b0a1662f69e req-a00e1c23-10ee-4c76-927f-a55dd9fd0eaf service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Received event network-vif-plugged-3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 632.207263] env[65121]: DEBUG oslo_concurrency.lockutils [req-687ab517-d6d2-43f5-8781-8b0a1662f69e req-a00e1c23-10ee-4c76-927f-a55dd9fd0eaf service nova] Acquiring lock "f1a5fdce-a940-43d4-83d7-716786c9ac34-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.207263] env[65121]: DEBUG oslo_concurrency.lockutils [req-687ab517-d6d2-43f5-8781-8b0a1662f69e req-a00e1c23-10ee-4c76-927f-a55dd9fd0eaf service nova] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.207376] env[65121]: DEBUG oslo_concurrency.lockutils [req-687ab517-d6d2-43f5-8781-8b0a1662f69e req-a00e1c23-10ee-4c76-927f-a55dd9fd0eaf service nova] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.207531] env[65121]: DEBUG nova.compute.manager [req-687ab517-d6d2-43f5-8781-8b0a1662f69e req-a00e1c23-10ee-4c76-927f-a55dd9fd0eaf service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] No waiting events found dispatching network-vif-plugged-3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 632.210422] env[65121]: WARNING nova.compute.manager [req-687ab517-d6d2-43f5-8781-8b0a1662f69e req-a00e1c23-10ee-4c76-927f-a55dd9fd0eaf service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Received unexpected event network-vif-plugged-3a66c1d6-9549-459d-bb1b-66781725a57f for instance with vm_state building and task_state spawning. [ 632.211020] env[65121]: DEBUG nova.compute.utils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 632.212660] env[65121]: DEBUG nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 632.213049] env[65121]: DEBUG nova.network.neutron [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 632.215668] env[65121]: WARNING neutronclient.v2_0.client [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.217254] env[65121]: WARNING neutronclient.v2_0.client [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.219153] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.219153] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.233735] env[65121]: INFO nova.compute.claims [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.252284] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105979, 'name': ReconfigVM_Task, 'duration_secs': 0.782501} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.252553] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Reconfigured VM instance instance-0000000a to attach disk [datastore1] cc002961-d742-4255-88c7-f0a5a39424b1/cc002961-d742-4255-88c7-f0a5a39424b1.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 632.253288] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40059904-b739-46a6-bf90-5c508af6f162 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.261850] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 632.261850] env[65121]: value = "task-5105983" [ 632.261850] env[65121]: _type = "Task" [ 632.261850] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.273756] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105983, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.338179] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105981, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.553646] env[65121]: DEBUG nova.policy [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f682931c5c9a406db2d6b1d000cfeb29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ca984047abf4272b00b44b551249fa1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 632.587877] env[65121]: DEBUG nova.network.neutron [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.624099] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105982, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.709151] env[65121]: DEBUG nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 632.740355] env[65121]: DEBUG nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 632.745190] env[65121]: INFO nova.compute.resource_tracker [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating resource usage from migration 5a7d1565-69d9-4c80-b059-fdb739cebe0a [ 632.776955] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105983, 'name': Rename_Task, 'duration_secs': 0.175778} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.777231] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 632.777473] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8092b2e2-4ba1-4da4-91b1-7e2c19ee8fb8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.787943] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 632.787943] env[65121]: value = "task-5105984" [ 632.787943] env[65121]: _type = "Task" [ 632.787943] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.797812] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105984, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.831674] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.831903] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.832093] env[65121]: DEBUG nova.compute.manager [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 632.832372] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105981, 'name': ReconfigVM_Task, 'duration_secs': 0.606727} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.835700] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b7120c-0dc9-4a4a-9082-1b00e1561630 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.838351] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Reconfigured VM instance instance-0000000b to attach disk [datastore1] f1a5fdce-a940-43d4-83d7-716786c9ac34/f1a5fdce-a940-43d4-83d7-716786c9ac34.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 632.839217] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b697b950-f481-48b5-81cb-9c4e20c064e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.845136] env[65121]: DEBUG nova.compute.manager [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 632.845897] env[65121]: DEBUG nova.objects.instance [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lazy-loading 'flavor' on Instance uuid 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 632.850305] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 632.850305] env[65121]: value = "task-5105985" [ 632.850305] env[65121]: _type = "Task" [ 632.850305] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.861333] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105985, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.090808] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.091336] env[65121]: DEBUG nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Instance network_info: |[{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 633.093047] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:95:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6966f473-59ac-49bb-9b7a-22c61f4e61e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ea2aa90-9370-491b-97b5-2c112a535219', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 633.101289] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Creating folder: Project (72237a0d762645588c41231b0a34a796). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 633.105048] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0276899b-54ce-46d9-bf92-9d319905d9e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.121266] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Created folder: Project (72237a0d762645588c41231b0a34a796) in parent group-v993268. [ 633.121266] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Creating folder: Instances. Parent ref: group-v993304. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 633.124714] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29579d4c-e447-47ea-971f-c303bb0dc78e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.130449] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105982, 'name': Rename_Task, 'duration_secs': 0.982254} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.130710] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 633.130982] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa23b82d-825b-4863-bf76-6cf54d786f20 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.139912] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 633.139912] env[65121]: value = "task-5105988" [ 633.139912] env[65121]: _type = "Task" [ 633.139912] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.140550] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22601d03-803e-4d85-90a7-c34f8e320c6c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.150353] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Created folder: Instances in parent group-v993304. [ 633.151068] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 633.154116] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 633.154472] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c718e6a-f17b-4655-b508-edb5b403216c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.173574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e672169-f484-44e7-9684-931327e87e3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.180878] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105988, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.213410] env[65121]: DEBUG nova.network.neutron [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Successfully created port: 16314fe2-22b5-437b-aebf-58bf042902d0 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 633.217759] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf07dfcb-4a9e-4463-afeb-9c030d7d929b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.222774] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 633.222774] env[65121]: value = "task-5105989" [ 633.222774] env[65121]: _type = "Task" [ 633.222774] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.234028] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6cddca-824d-47e1-9526-8bed39659fa8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.243625] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105989, 'name': CreateVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.251024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.260596] env[65121]: DEBUG nova.compute.provider_tree [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.301793] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105984, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.370588] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105985, 'name': Rename_Task, 'duration_secs': 0.206035} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.373584] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 633.374139] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf70ca24-3f40-4734-b4d2-be50912a4901 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.383152] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 633.383152] env[65121]: value = "task-5105990" [ 633.383152] env[65121]: _type = "Task" [ 633.383152] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.393071] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105990, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.656730] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105988, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.686813] env[65121]: DEBUG nova.network.neutron [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Successfully updated port: ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 633.737037] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105989, 'name': CreateVM_Task, 'duration_secs': 0.509896} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.737037] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 633.737037] env[65121]: WARNING neutronclient.v2_0.client [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.737037] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.737037] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.737527] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 633.738094] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-000f0558-f3b9-4ed0-9907-17102b1bc7a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.744199] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 633.744199] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52aaa580-08b8-7ae1-9a03-d9d77584ed2c" [ 633.744199] env[65121]: _type = "Task" [ 633.744199] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.758024] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52aaa580-08b8-7ae1-9a03-d9d77584ed2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.763503] env[65121]: DEBUG nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 633.766785] env[65121]: DEBUG nova.scheduler.client.report [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 633.796593] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 633.796785] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 633.797073] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 633.797321] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 633.797404] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 633.797524] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 633.797722] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.797892] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 633.798115] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 633.798363] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 633.798471] env[65121]: DEBUG nova.virt.hardware [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 633.800076] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d4c0d4-05fc-4bba-ae7e-fe79be0fee09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.806582] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105984, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.813829] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd69dc5b-4cb4-4fc4-98ce-873acf21121a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.872549] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 633.876204] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f2153cb-3483-40e2-bd92-a43c7a21e5f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.882957] env[65121]: DEBUG oslo_vmware.api [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 633.882957] env[65121]: value = "task-5105991" [ 633.882957] env[65121]: _type = "Task" [ 633.882957] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.904048] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105990, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.908692] env[65121]: DEBUG oslo_vmware.api [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105991, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.156854] env[65121]: DEBUG oslo_vmware.api [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5105988, 'name': PowerOnVM_Task, 'duration_secs': 0.680055} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.159598] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.159598] env[65121]: INFO nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Took 7.95 seconds to spawn the instance on the hypervisor. [ 634.159598] env[65121]: DEBUG nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.159598] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22eb5d7f-ad8d-4533-8001-030853c81d15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.189303] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.189584] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquired lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.191163] env[65121]: DEBUG nova.network.neutron [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 634.255549] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52aaa580-08b8-7ae1-9a03-d9d77584ed2c, 'name': SearchDatastore_Task, 'duration_secs': 0.01159} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.255863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.256172] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 634.256467] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.256656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.256896] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 634.257187] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c7a1e76-83aa-47fd-84d2-d348c0bdba7d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.269163] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 634.269359] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 634.271192] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e9dce3e-8b9a-431a-a03e-d7c456c2f85e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.275357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.566s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 634.275357] env[65121]: INFO nova.compute.manager [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Migrating [ 634.275357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.275487] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "compute-rpcapi-router" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.277283] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.678s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 634.277656] env[65121]: DEBUG nova.objects.instance [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Lazy-loading 'resources' on Instance uuid db737254-91fd-44bc-bf01-c864227cdf07 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 634.283333] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 634.283333] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525b4e42-7cf3-2761-d91c-5898e2a4a335" [ 634.283333] env[65121]: _type = "Task" [ 634.283333] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.296930] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525b4e42-7cf3-2761-d91c-5898e2a4a335, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.303208] env[65121]: DEBUG oslo_vmware.api [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5105984, 'name': PowerOnVM_Task, 'duration_secs': 1.167269} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.303542] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.303742] env[65121]: INFO nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Took 13.07 seconds to spawn the instance on the hypervisor. [ 634.303922] env[65121]: DEBUG nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.304829] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83cab2d-18ed-4e2d-ba5c-136c3f948bae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.396906] env[65121]: DEBUG oslo_vmware.api [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5105991, 'name': PowerOffVM_Task, 'duration_secs': 0.292269} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.397209] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 634.397454] env[65121]: DEBUG nova.compute.manager [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.398993] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e901f902-6432-4d12-ab1f-d06ccc78a4d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.406155] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105990, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.678084] env[65121]: INFO nova.compute.manager [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Took 27.46 seconds to build instance. [ 634.695141] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.695637] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.740643] env[65121]: DEBUG nova.network.neutron [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 634.762329] env[65121]: DEBUG nova.compute.manager [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received event network-vif-plugged-13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 634.762473] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.762698] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 634.762863] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 634.763034] env[65121]: DEBUG nova.compute.manager [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] No waiting events found dispatching network-vif-plugged-13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 634.763196] env[65121]: WARNING nova.compute.manager [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received unexpected event network-vif-plugged-13f8bd13-f8f0-4818-8dd6-845e9644a395 for instance with vm_state active and task_state None. [ 634.763347] env[65121]: DEBUG nova.compute.manager [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received event network-changed-13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 634.763562] env[65121]: DEBUG nova.compute.manager [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Refreshing instance network info cache due to event network-changed-13f8bd13-f8f0-4818-8dd6-845e9644a395. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 634.763748] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Acquiring lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.763895] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Acquired lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.764131] env[65121]: DEBUG nova.network.neutron [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Refreshing network info cache for port 13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 634.766483] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.766842] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.778606] env[65121]: INFO nova.compute.rpcapi [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Automatically selected compute RPC version 6.4 from minimum service version 70 [ 634.779375] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "compute-rpcapi-router" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.826737] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525b4e42-7cf3-2761-d91c-5898e2a4a335, 'name': SearchDatastore_Task, 'duration_secs': 0.022088} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.829368] env[65121]: INFO nova.compute.manager [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Took 36.21 seconds to build instance. [ 634.839101] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b57f70d6-4194-4333-a59a-f0ca27a96b4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.849950] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 634.849950] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528334bc-afbf-f2b3-3a4e-c47eccc30fe2" [ 634.849950] env[65121]: _type = "Task" [ 634.849950] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.861871] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528334bc-afbf-f2b3-3a4e-c47eccc30fe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.905232] env[65121]: DEBUG oslo_vmware.api [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5105990, 'name': PowerOnVM_Task, 'duration_secs': 1.341211} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.907047] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.908341] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.909665] env[65121]: WARNING openstack [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.929162] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.929162] env[65121]: INFO nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Took 11.20 seconds to spawn the instance on the hypervisor. [ 634.929162] env[65121]: DEBUG nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.933455] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921de5a1-7691-4898-9c1c-5ebe26082e31 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.937729] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ce61af4b-933a-41e4-b8a8-0c4a58a52840 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.105s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.180247] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cfa10c41-32f8-419a-b4b6-0214999b21fe tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "6887bb66-7186-4d58-be59-d0c53c3221c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.145s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.200058] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b71373-3bac-4239-bc31-4258240c06e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.211791] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d53668d-1c83-41e6-9124-4c1a2c56a6f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.253169] env[65121]: DEBUG nova.network.neutron [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Successfully updated port: 16314fe2-22b5-437b-aebf-58bf042902d0 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 635.256517] env[65121]: DEBUG nova.network.neutron [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Updating instance_info_cache with network_info: [{"id": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "address": "fa:16:3e:6b:03:cf", "network": {"id": "81cf3ab5-bc0a-4335-88c1-87fc12584925", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-98268708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "def20b65ac8341c1a60a747260589d5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec7592a3-0a", "ovs_interfaceid": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 635.258161] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb6d7ea-6d2d-405b-b7d0-69debd4a0dc7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.269265] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7352b5b2-60d4-4d38-8153-a74b123d1930 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.283309] env[65121]: WARNING neutronclient.v2_0.client [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.283993] env[65121]: WARNING openstack [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.284343] env[65121]: WARNING openstack [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.291528] env[65121]: DEBUG nova.compute.provider_tree [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.308377] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.308377] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.308377] env[65121]: DEBUG nova.network.neutron [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 635.339639] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb675361-fb3a-44fb-9730-d269634470c9 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "cc002961-d742-4255-88c7-f0a5a39424b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.733s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.344374] env[65121]: INFO nova.compute.manager [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Rebuilding instance [ 635.360698] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528334bc-afbf-f2b3-3a4e-c47eccc30fe2, 'name': SearchDatastore_Task, 'duration_secs': 0.036568} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.360958] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.361223] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] c4096314-270a-4270-9e1d-5ace8ddbd286/c4096314-270a-4270-9e1d-5ace8ddbd286.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 635.361506] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40fc13e2-e0ee-4b5a-ae60-6e7da67d3e54 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.370257] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 635.370257] env[65121]: value = "task-5105992" [ 635.370257] env[65121]: _type = "Task" [ 635.370257] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.390930] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105992, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.398838] env[65121]: DEBUG nova.compute.manager [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 635.399712] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3e4003-87ba-4b65-8705-77908eec20c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.412743] env[65121]: WARNING openstack [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.413193] env[65121]: WARNING openstack [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.458134] env[65121]: INFO nova.compute.manager [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Took 28.53 seconds to build instance. [ 635.488888] env[65121]: WARNING neutronclient.v2_0.client [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.489586] env[65121]: WARNING openstack [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.489978] env[65121]: WARNING openstack [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.581037] env[65121]: DEBUG nova.network.neutron [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updated VIF entry in instance network info cache for port 13f8bd13-f8f0-4818-8dd6-845e9644a395. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 635.581412] env[65121]: DEBUG nova.network.neutron [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 635.762551] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "refresh_cache-46f89cfb-14aa-483d-aa8a-bc91356272ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.762864] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquired lock "refresh_cache-46f89cfb-14aa-483d-aa8a-bc91356272ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.763085] env[65121]: DEBUG nova.network.neutron [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 635.764660] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Releasing lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.764987] env[65121]: DEBUG nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Instance network_info: |[{"id": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "address": "fa:16:3e:6b:03:cf", "network": {"id": "81cf3ab5-bc0a-4335-88c1-87fc12584925", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-98268708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "def20b65ac8341c1a60a747260589d5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec7592a3-0a", "ovs_interfaceid": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 635.767046] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:03:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec7592a3-0a2f-43b5-9c55-2e59d85c17cf', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 635.777234] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Creating folder: Project (def20b65ac8341c1a60a747260589d5a). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 635.778560] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4c0f623-6b3c-4bcf-8867-b90109b8704d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.791129] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Created folder: Project (def20b65ac8341c1a60a747260589d5a) in parent group-v993268. [ 635.791336] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Creating folder: Instances. Parent ref: group-v993307. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 635.791610] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a02482d-9c12-4572-94be-c36ea0b00468 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.794703] env[65121]: DEBUG nova.scheduler.client.report [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 635.813170] env[65121]: WARNING neutronclient.v2_0.client [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.813823] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.814184] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.821565] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Created folder: Instances in parent group-v993307. [ 635.821835] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 635.823592] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 635.824724] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48a7a0a9-5995-45bc-8822-f52e438e6cc8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.849344] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 635.849344] env[65121]: value = "task-5105995" [ 635.849344] env[65121]: _type = "Task" [ 635.849344] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.864194] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105995, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.885107] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105992, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.948449] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.949188] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.961167] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc6a0683-4a3e-4ac0-9154-723e3c9bd7d3 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.701s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.048078] env[65121]: WARNING neutronclient.v2_0.client [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.049250] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.053048] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.084028] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a583fde-6c6a-4149-a51a-82ad69e94661 req-e6985db7-3b48-4403-9fd5-8c04d52294d0 service nova] Releasing lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.164113] env[65121]: DEBUG nova.network.neutron [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance_info_cache with network_info: [{"id": "fc77c1b5-9d9b-479a-a553-e7961313af08", "address": "fa:16:3e:f6:e5:10", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.52", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc77c1b5-9d", "ovs_interfaceid": "fc77c1b5-9d9b-479a-a553-e7961313af08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 636.268913] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.269643] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.301645] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.024s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.304266] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.170s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.304513] env[65121]: DEBUG nova.objects.instance [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lazy-loading 'resources' on Instance uuid 0151ed85-bba4-443a-b18f-58272c973e1e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 636.310477] env[65121]: DEBUG nova.network.neutron [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 636.323726] env[65121]: INFO nova.scheduler.client.report [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Deleted allocations for instance db737254-91fd-44bc-bf01-c864227cdf07 [ 636.354552] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.354993] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.378993] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5105995, 'name': CreateVM_Task, 'duration_secs': 0.408863} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.379859] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 636.380367] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.380705] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.380853] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.381425] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 636.384906] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44887ba1-e9b2-48e3-a876-5bb2da0672d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.386648] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105992, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672054} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.386967] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] c4096314-270a-4270-9e1d-5ace8ddbd286/c4096314-270a-4270-9e1d-5ace8ddbd286.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 636.387161] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 636.391068] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a067a1f4-fa8b-49c5-ba55-82c29823542d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.391232] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 636.391232] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5218bff3-851b-ab14-9e2d-f91985eab8d8" [ 636.391232] env[65121]: _type = "Task" [ 636.391232] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.403238] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 636.403238] env[65121]: value = "task-5105996" [ 636.403238] env[65121]: _type = "Task" [ 636.403238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.411514] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5218bff3-851b-ab14-9e2d-f91985eab8d8, 'name': SearchDatastore_Task, 'duration_secs': 0.013554} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.412343] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.412705] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 636.412963] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.413161] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.413359] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 636.413680] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad3c29cb-e467-4a2f-8e5e-12d240ead0fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.420249] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105996, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.421342] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 636.421614] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2410ee45-41d3-4bec-abc5-8d969f819075 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.429605] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 636.429811] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 636.432342] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d6a570f-50c9-465f-9817-66d59ac6926b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.435074] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 636.435074] env[65121]: value = "task-5105997" [ 636.435074] env[65121]: _type = "Task" [ 636.435074] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.441927] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 636.441927] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5299dbd2-a483-bad1-a1f5-f0f19fba26c3" [ 636.441927] env[65121]: _type = "Task" [ 636.441927] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.448768] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5105997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.456692] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5299dbd2-a483-bad1-a1f5-f0f19fba26c3, 'name': SearchDatastore_Task, 'duration_secs': 0.013679} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.457957] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fe4b542-e822-4de0-8cd9-c063d4e22ca7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.464622] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 636.464622] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529a0887-6caf-795f-c851-a55ab01e0707" [ 636.464622] env[65121]: _type = "Task" [ 636.464622] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.473922] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529a0887-6caf-795f-c851-a55ab01e0707, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.475215] env[65121]: WARNING neutronclient.v2_0.client [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.475965] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.476334] env[65121]: WARNING openstack [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.534461] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "6218957b-6329-4004-97ca-07231b55937a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.535295] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "6218957b-6329-4004-97ca-07231b55937a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.538255] env[65121]: DEBUG nova.compute.manager [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Received event network-changed-3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 636.538255] env[65121]: DEBUG nova.compute.manager [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Refreshing instance network info cache due to event network-changed-3a66c1d6-9549-459d-bb1b-66781725a57f. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 636.538584] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Acquiring lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.538672] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Acquired lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.538864] env[65121]: DEBUG nova.network.neutron [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Refreshing network info cache for port 3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 636.594713] env[65121]: DEBUG nova.network.neutron [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Updating instance_info_cache with network_info: [{"id": "16314fe2-22b5-437b-aebf-58bf042902d0", "address": "fa:16:3e:30:56:f3", "network": {"id": "753fce83-9625-407b-ad4a-3a8c260fad09", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-642414393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ca984047abf4272b00b44b551249fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16314fe2-22", "ovs_interfaceid": "16314fe2-22b5-437b-aebf-58bf042902d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 636.669745] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.831840] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8f1f5a54-3d66-4ee5-b6a5-6ff2bbb8b6ed tempest-DeleteServersAdminTestJSON-1986721668 tempest-DeleteServersAdminTestJSON-1986721668-project-admin] Lock "db737254-91fd-44bc-bf01-c864227cdf07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.614s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.915523] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105996, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085374} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.918670] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 636.920326] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056a7336-9a42-451f-a300-d7b595823dde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.945488] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] c4096314-270a-4270-9e1d-5ace8ddbd286/c4096314-270a-4270-9e1d-5ace8ddbd286.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 636.951360] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04b54ab1-1bbf-426f-ac4a-2001c53b4586 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.974880] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5105997, 'name': PowerOffVM_Task, 'duration_secs': 0.161327} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.975618] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 636.975914] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.977848] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f136c9b-a2d5-4ac3-bb28-dccc48f2ae33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.980910] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 636.980910] env[65121]: value = "task-5105998" [ 636.980910] env[65121]: _type = "Task" [ 636.980910] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.987055] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529a0887-6caf-795f-c851-a55ab01e0707, 'name': SearchDatastore_Task, 'duration_secs': 0.012068} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.990847] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.991115] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818/4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.993363] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dfe007e0-d306-4bab-adb4-729d51d3d9e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.995701] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 636.996282] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3f729a6-814b-4a9f-92ee-81e7cbc21e76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.004658] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.006086] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 637.006086] env[65121]: value = "task-5105999" [ 637.006086] env[65121]: _type = "Task" [ 637.006086] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.015942] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5105999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.042446] env[65121]: DEBUG nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 637.044266] env[65121]: WARNING neutronclient.v2_0.client [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.044935] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.045476] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.052175] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 637.052381] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 637.052555] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Deleting the datastore file [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 637.056094] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-baf65fdc-e9f2-4be7-8e5c-d0ac1049d1df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.067927] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 637.067927] env[65121]: value = "task-5106001" [ 637.067927] env[65121]: _type = "Task" [ 637.067927] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.078038] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.097854] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Releasing lock "refresh_cache-46f89cfb-14aa-483d-aa8a-bc91356272ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.098304] env[65121]: DEBUG nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Instance network_info: |[{"id": "16314fe2-22b5-437b-aebf-58bf042902d0", "address": "fa:16:3e:30:56:f3", "network": {"id": "753fce83-9625-407b-ad4a-3a8c260fad09", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-642414393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ca984047abf4272b00b44b551249fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16314fe2-22", "ovs_interfaceid": "16314fe2-22b5-437b-aebf-58bf042902d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 637.098844] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:56:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '16314fe2-22b5-437b-aebf-58bf042902d0', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 637.107948] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Creating folder: Project (0ca984047abf4272b00b44b551249fa1). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.108481] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-115f9cb8-435e-4e56-9e03-808437f3d84f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.120572] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Created folder: Project (0ca984047abf4272b00b44b551249fa1) in parent group-v993268. [ 637.120798] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Creating folder: Instances. Parent ref: group-v993310. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.121502] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53200c19-318c-4c01-b2f3-9cd956825eca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.134688] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Created folder: Instances in parent group-v993310. [ 637.134947] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 637.135194] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 637.136388] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ac5dea7-a65d-4d57-b84b-c3586cd565f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.169018] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 637.169018] env[65121]: value = "task-5106004" [ 637.169018] env[65121]: _type = "Task" [ 637.169018] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.189832] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106004, 'name': CreateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.223815] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.224401] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.273648] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f244846-4276-4252-a669-bff4b1488541 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.285894] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9100457-2575-4e28-8fc2-a9be13d11604 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.331611] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e833e2a5-145c-4698-a5d8-6123c430e25e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.335326] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.335606] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.344768] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd46499-f8ce-41fc-aa88-9ae1376d810d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.364321] env[65121]: DEBUG nova.compute.provider_tree [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.368191] env[65121]: WARNING neutronclient.v2_0.client [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.368503] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.369048] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.383419] env[65121]: DEBUG nova.scheduler.client.report [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 637.500258] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.517703] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5105999, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.520092] env[65121]: DEBUG nova.network.neutron [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updated VIF entry in instance network info cache for port 3a66c1d6-9549-459d-bb1b-66781725a57f. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 637.520463] env[65121]: DEBUG nova.network.neutron [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updating instance_info_cache with network_info: [{"id": "3a66c1d6-9549-459d-bb1b-66781725a57f", "address": "fa:16:3e:be:bf:37", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66c1d6-95", "ovs_interfaceid": "3a66c1d6-9549-459d-bb1b-66781725a57f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 637.581262] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.586053] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106001, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242051} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.586245] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 637.586915] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 637.586915] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.662812] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.663200] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.663466] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.663712] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.663884] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 637.666940] env[65121]: INFO nova.compute.manager [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Terminating instance [ 637.689192] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106004, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.703327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "61bd5995-701e-430f-9aae-4b266089e313" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.703610] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.711920] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "4b6fe349-335e-4202-ab76-04f6ccc036ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.712167] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.842185] env[65121]: DEBUG nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 637.889402] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.585s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 637.893078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.485s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.895043] env[65121]: INFO nova.compute.claims [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.933581] env[65121]: INFO nova.scheduler.client.report [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Deleted allocations for instance 0151ed85-bba4-443a-b18f-58272c973e1e [ 638.003677] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5105998, 'name': ReconfigVM_Task, 'duration_secs': 0.587628} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.003677] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Reconfigured VM instance instance-0000000d to attach disk [datastore1] c4096314-270a-4270-9e1d-5ace8ddbd286/c4096314-270a-4270-9e1d-5ace8ddbd286.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 638.005134] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa2e9ed1-9e0a-4c09-95ec-864f26e3dc3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.019620] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5105999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.754173} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.021221] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818/4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 638.021754] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 638.021904] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 638.021904] env[65121]: value = "task-5106005" [ 638.021904] env[65121]: _type = "Task" [ 638.021904] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.022836] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-622935c3-4b2e-4b87-9615-03687e22fe45 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.025246] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Releasing lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.025327] env[65121]: DEBUG nova.compute.manager [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Received event network-vif-plugged-4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 638.025503] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.025703] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.025857] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.026022] env[65121]: DEBUG nova.compute.manager [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] No waiting events found dispatching network-vif-plugged-4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 638.026183] env[65121]: WARNING nova.compute.manager [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Received unexpected event network-vif-plugged-4ea2aa90-9370-491b-97b5-2c112a535219 for instance with vm_state building and task_state spawning. [ 638.026334] env[65121]: DEBUG nova.compute.manager [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Received event network-changed-4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 638.027024] env[65121]: DEBUG nova.compute.manager [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Refreshing instance network info cache due to event network-changed-4ea2aa90-9370-491b-97b5-2c112a535219. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 638.027024] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.027024] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.027024] env[65121]: DEBUG nova.network.neutron [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Refreshing network info cache for port 4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 638.039105] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106005, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.040616] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 638.040616] env[65121]: value = "task-5106006" [ 638.040616] env[65121]: _type = "Task" [ 638.040616] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.177210] env[65121]: DEBUG nova.compute.manager [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 638.177741] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.181602] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180ffa98-8300-4d2a-800d-4ca2eeb2ecf2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.185115] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106004, 'name': CreateVM_Task, 'duration_secs': 0.794622} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.185294] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 638.186202] env[65121]: WARNING neutronclient.v2_0.client [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.186586] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.186730] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.187054] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 638.187318] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-253f018e-b5e6-468b-90ee-37b55bbb2b08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.193249] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 638.193951] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1692876f-a3f9-4b9a-b3a8-fb3031137484 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.198428] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 638.198428] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521942c5-9b66-5508-a37e-1bb8e589ac8d" [ 638.198428] env[65121]: _type = "Task" [ 638.198428] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.199368] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9ce072-5662-43b1-b12b-9071e5344296 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.204631] env[65121]: DEBUG oslo_vmware.api [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 638.204631] env[65121]: value = "task-5106007" [ 638.204631] env[65121]: _type = "Task" [ 638.204631] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.209118] env[65121]: DEBUG nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 638.236204] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance '5f6e4f46-0745-42c0-a779-6cffb60e21ca' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 638.244023] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521942c5-9b66-5508-a37e-1bb8e589ac8d, 'name': SearchDatastore_Task, 'duration_secs': 0.020425} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.244667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.244898] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 638.245142] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.245275] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.245437] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 638.245776] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7693cd3-a991-4d4a-a594-17a1cde2b51f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.251127] env[65121]: DEBUG oslo_vmware.api [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106007, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.260280] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 638.260504] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 638.261568] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1030879-4a54-4229-8832-a1d45f7894e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.269728] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 638.269728] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524ac58d-6055-4f81-f53c-4c9623e9e134" [ 638.269728] env[65121]: _type = "Task" [ 638.269728] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.281117] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524ac58d-6055-4f81-f53c-4c9623e9e134, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.369810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.443999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f5f623b-25ad-4766-acd6-a7b2cfe81c18 tempest-ServerDiagnosticsNegativeTest-142732464 tempest-ServerDiagnosticsNegativeTest-142732464-project-member] Lock "0151ed85-bba4-443a-b18f-58272c973e1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.874s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.466998] env[65121]: DEBUG nova.compute.manager [None req-1d028a8d-8be6-4a8d-867b-b3163b519c04 tempest-ServerDiagnosticsV248Test-217477461 tempest-ServerDiagnosticsV248Test-217477461-project-admin] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 638.468386] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9606d18-25ca-442d-9658-e92052780d8f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.478098] env[65121]: INFO nova.compute.manager [None req-1d028a8d-8be6-4a8d-867b-b3163b519c04 tempest-ServerDiagnosticsV248Test-217477461 tempest-ServerDiagnosticsV248Test-217477461-project-admin] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Retrieving diagnostics [ 638.478707] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f9f3ce-e4cc-4a1c-96dd-5f285b8e9c75 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.532321] env[65121]: WARNING neutronclient.v2_0.client [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.533018] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.533817] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.544561] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106005, 'name': Rename_Task, 'duration_secs': 0.334378} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.548066] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 638.548639] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26b4a60d-d4a1-43f3-82c2-7f8d00327d5d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.557187] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106006, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.203284} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.559057] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 638.559207] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 638.559207] env[65121]: value = "task-5106008" [ 638.559207] env[65121]: _type = "Task" [ 638.559207] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.559984] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c25c8b3-8633-46b3-8243-656e049d4a7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.590770] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818/4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 638.593811] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-704eecf6-28c6-416c-8e72-9f4bb42ee267 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.621578] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 638.621578] env[65121]: value = "task-5106009" [ 638.621578] env[65121]: _type = "Task" [ 638.621578] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.631960] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106009, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.647268] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 638.647418] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 638.647832] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 638.647916] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 638.648073] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 638.648229] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 638.648480] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.648672] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 638.648846] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 638.649015] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 638.649188] env[65121]: DEBUG nova.virt.hardware [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 638.650099] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3906bd4f-2bbb-4c80-962a-bd959c846d01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.665348] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47517969-092a-4e1a-a340-9db5af924cb9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.680281] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 638.685985] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 638.686257] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 638.686380] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ac0ca3b-be6f-4e3e-943c-4f392d76a63b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.699464] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.699778] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.714103] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 638.714103] env[65121]: value = "task-5106010" [ 638.714103] env[65121]: _type = "Task" [ 638.714103] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.724789] env[65121]: DEBUG oslo_vmware.api [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106007, 'name': PowerOffVM_Task, 'duration_secs': 0.332897} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.725560] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 638.725937] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 638.728234] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b01ef015-051b-4cd2-b070-b1872e4c513e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.733182] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106010, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.743706] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.747234] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 638.747234] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5a13d5a-8732-4318-b5e1-50a87821b5d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.759315] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 638.759315] env[65121]: value = "task-5106012" [ 638.759315] env[65121]: _type = "Task" [ 638.759315] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.772297] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.781666] env[65121]: WARNING neutronclient.v2_0.client [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.782382] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.782750] env[65121]: WARNING openstack [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.797295] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524ac58d-6055-4f81-f53c-4c9623e9e134, 'name': SearchDatastore_Task, 'duration_secs': 0.020233} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.799039] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-007979dd-cbd8-4ba5-b2d8-ef17c56ab2f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.808577] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 638.808577] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5290a2e3-b95c-23cf-a8ae-ccc3fa27610d" [ 638.808577] env[65121]: _type = "Task" [ 638.808577] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.819400] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5290a2e3-b95c-23cf-a8ae-ccc3fa27610d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.834824] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 638.835145] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 638.835386] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleting the datastore file [datastore2] bad21c0b-6534-46b5-97a3-a26e5ecb5c96 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 638.835719] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b118769-dd86-4705-8c16-4414f4faa72a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.845938] env[65121]: DEBUG oslo_vmware.api [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 638.845938] env[65121]: value = "task-5106013" [ 638.845938] env[65121]: _type = "Task" [ 638.845938] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.855589] env[65121]: DEBUG oslo_vmware.api [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.888919] env[65121]: DEBUG nova.network.neutron [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updated VIF entry in instance network info cache for port 4ea2aa90-9370-491b-97b5-2c112a535219. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 638.889258] env[65121]: DEBUG nova.network.neutron [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.079702] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106008, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.135768] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106009, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.228052] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106010, 'name': CreateVM_Task, 'duration_secs': 0.442914} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.228052] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 639.228052] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.228052] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.228052] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 639.228263] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebe5dc04-53f3-4b79-90c2-bcc51f426c1c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.234695] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 639.234695] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d9af2f-d9f8-74ce-6c74-a802d2491ae9" [ 639.234695] env[65121]: _type = "Task" [ 639.234695] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.247957] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d9af2f-d9f8-74ce-6c74-a802d2491ae9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.271488] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106012, 'name': PowerOffVM_Task, 'duration_secs': 0.322416} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.271804] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 639.273221] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance '5f6e4f46-0745-42c0-a779-6cffb60e21ca' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 639.332539] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5290a2e3-b95c-23cf-a8ae-ccc3fa27610d, 'name': SearchDatastore_Task, 'duration_secs': 0.016063} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.332992] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.334583] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 46f89cfb-14aa-483d-aa8a-bc91356272ca/46f89cfb-14aa-483d-aa8a-bc91356272ca.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 639.334583] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14b9c991-3868-4fcf-b39a-c6c4c6eafd43 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.340497] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6471123c-21d3-4dec-a864-7ca7a2424217 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.353405] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 639.353405] env[65121]: value = "task-5106014" [ 639.353405] env[65121]: _type = "Task" [ 639.353405] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.355187] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3222fbcb-2f96-4e90-9173-d2f737032004 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.364973] env[65121]: DEBUG oslo_vmware.api [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258304} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.365759] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 639.365824] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 639.366017] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.366225] env[65121]: INFO nova.compute.manager [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Took 1.19 seconds to destroy the instance on the hypervisor. [ 639.366434] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 639.366935] env[65121]: DEBUG nova.compute.manager [-] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 639.367049] env[65121]: DEBUG nova.network.neutron [-] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 639.367300] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.367841] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.368120] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.408477] env[65121]: DEBUG oslo_concurrency.lockutils [req-587f3433-81f5-4241-8ba2-df9ea34969a9 req-aef4e9cb-6046-42d3-a57d-02aa61870db2 service nova] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.408990] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106014, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.410602] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Received event network-vif-plugged-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 639.410811] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Acquiring lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.411626] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.411626] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.411626] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] No waiting events found dispatching network-vif-plugged-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 639.411626] env[65121]: WARNING nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Received unexpected event network-vif-plugged-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf for instance with vm_state building and task_state spawning. [ 639.412195] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Received event network-changed-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 639.412195] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Refreshing instance network info cache due to event network-changed-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 639.412195] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Acquiring lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.412195] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Acquired lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.412474] env[65121]: DEBUG nova.network.neutron [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Refreshing network info cache for port ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 639.414289] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed46ebf7-61f7-4215-a06a-16afcf35d4a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.430508] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90180a28-0ede-4b3f-b6ad-883dbad5e3d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.446430] env[65121]: DEBUG nova.compute.provider_tree [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.556851] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.577397] env[65121]: DEBUG oslo_vmware.api [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106008, 'name': PowerOnVM_Task, 'duration_secs': 1.000394} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.577725] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 639.578038] env[65121]: INFO nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Took 10.91 seconds to spawn the instance on the hypervisor. [ 639.578269] env[65121]: DEBUG nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 639.579335] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ea78fa-e2f9-4e4c-bd94-d795e7c2cc05 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.634772] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106009, 'name': ReconfigVM_Task, 'duration_secs': 0.718864} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.636187] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Reconfigured VM instance instance-0000000e to attach disk [datastore2] 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818/4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.636187] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc2d3636-a8c2-425f-9421-8e9d7546d100 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.644799] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 639.644799] env[65121]: value = "task-5106015" [ 639.644799] env[65121]: _type = "Task" [ 639.644799] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.664629] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106015, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.752693] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d9af2f-d9f8-74ce-6c74-a802d2491ae9, 'name': SearchDatastore_Task, 'duration_secs': 0.048672} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.752863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.754370] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 639.754370] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.754370] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.754370] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 639.754370] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f60f0e1b-c06f-4df0-a3f2-fb66348a935e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.765822] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 639.765995] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 639.767043] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d676b02-0c18-4e19-b12d-3f84c31a0f09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.774927] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 639.774927] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5251eb73-ac1d-f321-f81b-dc42f78d8c7d" [ 639.774927] env[65121]: _type = "Task" [ 639.774927] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.779919] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 639.780129] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 639.780129] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 639.780310] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 639.780617] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 639.780617] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 639.780752] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.780901] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 639.781073] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 639.781225] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 639.781407] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 639.788199] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89d2f11d-608e-4a71-a581-8d1dcb49735c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.806800] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5251eb73-ac1d-f321-f81b-dc42f78d8c7d, 'name': SearchDatastore_Task, 'duration_secs': 0.017738} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.810209] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 639.810209] env[65121]: value = "task-5106016" [ 639.810209] env[65121]: _type = "Task" [ 639.810209] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.811110] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66f3fc45-747f-4248-a9b6-5e76ab7ef6da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.824442] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106016, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.828108] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 639.828108] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b45e90-df5f-d946-bbac-8cb1ef799dbb" [ 639.828108] env[65121]: _type = "Task" [ 639.828108] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.836518] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b45e90-df5f-d946-bbac-8cb1ef799dbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.865735] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106014, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.912515] env[65121]: DEBUG nova.compute.manager [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 639.912515] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b7af1f-25b6-4533-b253-6cd6f844669e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.918913] env[65121]: WARNING neutronclient.v2_0.client [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.919565] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.920042] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.951312] env[65121]: DEBUG nova.scheduler.client.report [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 640.104035] env[65121]: INFO nova.compute.manager [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Took 31.61 seconds to build instance. [ 640.158493] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106015, 'name': Rename_Task, 'duration_secs': 0.225283} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.159059] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "5e915c8e-150a-4bfd-8005-a7a1a119057d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.159157] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.159368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "5e915c8e-150a-4bfd-8005-a7a1a119057d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.159587] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.159755] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.161763] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 640.162613] env[65121]: INFO nova.compute.manager [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Terminating instance [ 640.164128] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61ba5f53-3632-435f-b0c7-f709380b9821 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.174751] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 640.174751] env[65121]: value = "task-5106017" [ 640.174751] env[65121]: _type = "Task" [ 640.174751] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.233334] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.233486] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.326992] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106016, 'name': ReconfigVM_Task, 'duration_secs': 0.268377} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.330781] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance '5f6e4f46-0745-42c0-a779-6cffb60e21ca' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 640.336212] env[65121]: WARNING neutronclient.v2_0.client [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.337587] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.337587] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.351711] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b45e90-df5f-d946-bbac-8cb1ef799dbb, 'name': SearchDatastore_Task, 'duration_secs': 0.084374} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.352036] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.352304] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 640.352592] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59896868-bba1-448c-a685-31c6db0296b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.361367] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 640.361367] env[65121]: value = "task-5106018" [ 640.361367] env[65121]: _type = "Task" [ 640.361367] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.364240] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106014, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.853394} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.367439] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 46f89cfb-14aa-483d-aa8a-bc91356272ca/46f89cfb-14aa-483d-aa8a-bc91356272ca.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 640.367652] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 640.368015] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4483f292-6b03-4c11-b462-050d8d94c3f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.376135] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106018, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.378417] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 640.378417] env[65121]: value = "task-5106019" [ 640.378417] env[65121]: _type = "Task" [ 640.378417] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.389736] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106019, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.439341] env[65121]: INFO nova.compute.manager [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] instance snapshotting [ 640.439595] env[65121]: WARNING nova.compute.manager [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 640.442320] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26e5e3a-e3ce-478d-a891-8803c91d918e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.463927] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.464537] env[65121]: DEBUG nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 640.470255] env[65121]: DEBUG nova.network.neutron [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Updated VIF entry in instance network info cache for port ec7592a3-0a2f-43b5-9c55-2e59d85c17cf. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 640.470584] env[65121]: DEBUG nova.network.neutron [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Updating instance_info_cache with network_info: [{"id": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "address": "fa:16:3e:6b:03:cf", "network": {"id": "81cf3ab5-bc0a-4335-88c1-87fc12584925", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-98268708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "def20b65ac8341c1a60a747260589d5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec7592a3-0a", "ovs_interfaceid": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 640.472094] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.461s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.473745] env[65121]: INFO nova.compute.claims [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.477081] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29163f7d-6d7e-4326-843d-e27e04400b3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.606368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8a98f7f-680b-414d-a50f-9dd3f6706a97 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.921s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.621127] env[65121]: DEBUG nova.network.neutron [-] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 640.670465] env[65121]: DEBUG nova.compute.manager [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 640.671017] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.671657] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1b0c24-3e94-43c7-9af0-389c041fa755 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.684829] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106017, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.686678] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 640.686872] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7b3a5be-1b50-4fd7-a6e1-9d5bdf1a32ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.697022] env[65121]: DEBUG oslo_vmware.api [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 640.697022] env[65121]: value = "task-5106020" [ 640.697022] env[65121]: _type = "Task" [ 640.697022] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.709722] env[65121]: DEBUG oslo_vmware.api [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.839261] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 640.839555] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 640.840268] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 640.840268] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 640.840268] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 640.840268] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 640.840533] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.841087] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 640.841087] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 640.841087] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 640.841087] env[65121]: DEBUG nova.virt.hardware [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 640.847284] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Reconfiguring VM instance instance-00000001 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 640.847709] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d56a04a-7e7c-4be3-acc8-65c6aed7a13a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.877611] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106018, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.879707] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 640.879707] env[65121]: value = "task-5106021" [ 640.879707] env[65121]: _type = "Task" [ 640.879707] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.895158] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106021, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.903116] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106019, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08791} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.903591] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.903852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.904579] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 640.906129] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38aaf63-362e-4e78-abb1-3b3ef43d4bd3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.940642] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 46f89cfb-14aa-483d-aa8a-bc91356272ca/46f89cfb-14aa-483d-aa8a-bc91356272ca.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 640.941069] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0b9c4e3-9273-4bd3-9926-3fbb744dcad6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.970774] env[65121]: DEBUG nova.compute.utils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 640.973344] env[65121]: DEBUG nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 640.974132] env[65121]: DEBUG nova.network.neutron [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 640.974483] env[65121]: WARNING neutronclient.v2_0.client [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.974948] env[65121]: WARNING neutronclient.v2_0.client [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.975884] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.976504] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.989345] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Releasing lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.989345] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Received event network-vif-plugged-16314fe2-22b5-437b-aebf-58bf042902d0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 640.989345] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Acquiring lock "46f89cfb-14aa-483d-aa8a-bc91356272ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.989345] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.989827] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.989827] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] No waiting events found dispatching network-vif-plugged-16314fe2-22b5-437b-aebf-58bf042902d0 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 640.989827] env[65121]: WARNING nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Received unexpected event network-vif-plugged-16314fe2-22b5-437b-aebf-58bf042902d0 for instance with vm_state building and task_state spawning. [ 640.989937] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Received event network-changed-16314fe2-22b5-437b-aebf-58bf042902d0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 640.990091] env[65121]: DEBUG nova.compute.manager [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Refreshing instance network info cache due to event network-changed-16314fe2-22b5-437b-aebf-58bf042902d0. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 640.990286] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Acquiring lock "refresh_cache-46f89cfb-14aa-483d-aa8a-bc91356272ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.990762] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Acquired lock "refresh_cache-46f89cfb-14aa-483d-aa8a-bc91356272ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.990762] env[65121]: DEBUG nova.network.neutron [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Refreshing network info cache for port 16314fe2-22b5-437b-aebf-58bf042902d0 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 640.998901] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 640.999298] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 640.999298] env[65121]: value = "task-5106022" [ 640.999298] env[65121]: _type = "Task" [ 640.999298] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.999651] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e635750d-59af-4d4e-95b8-a37d18d1df10 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.020597] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.023140] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 641.023140] env[65121]: value = "task-5106023" [ 641.023140] env[65121]: _type = "Task" [ 641.023140] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.040325] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106023, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.059641] env[65121]: DEBUG nova.compute.manager [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received event network-changed-13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 641.059641] env[65121]: DEBUG nova.compute.manager [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Refreshing instance network info cache due to event network-changed-13f8bd13-f8f0-4818-8dd6-845e9644a395. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 641.060830] env[65121]: DEBUG oslo_concurrency.lockutils [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Acquiring lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.060830] env[65121]: DEBUG oslo_concurrency.lockutils [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Acquired lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.060830] env[65121]: DEBUG nova.network.neutron [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Refreshing network info cache for port 13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 641.095942] env[65121]: DEBUG nova.policy [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fd28a50f09f48959706f1637ca9f3b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ede261894564e9d9bf50ce9cddeec77', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 641.110297] env[65121]: DEBUG nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 641.125356] env[65121]: INFO nova.compute.manager [-] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Took 1.76 seconds to deallocate network for instance. [ 641.188121] env[65121]: DEBUG oslo_vmware.api [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106017, 'name': PowerOnVM_Task, 'duration_secs': 0.976524} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.188804] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 641.189083] env[65121]: INFO nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Took 9.98 seconds to spawn the instance on the hypervisor. [ 641.189274] env[65121]: DEBUG nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 641.190129] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bd2544-e28a-4693-b889-33f0aad2e88a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.215286] env[65121]: DEBUG oslo_vmware.api [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106020, 'name': PowerOffVM_Task, 'duration_secs': 0.302085} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.215920] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 641.215920] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 641.216108] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7805a199-61cf-4a87-abf4-8a1ebffe30da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.314063] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 641.314312] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 641.314483] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleting the datastore file [datastore2] 5e915c8e-150a-4bfd-8005-a7a1a119057d {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 641.314859] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b28fdc58-abba-4de8-b7e4-ae3d541dbbf4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.326870] env[65121]: DEBUG oslo_vmware.api [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 641.326870] env[65121]: value = "task-5106025" [ 641.326870] env[65121]: _type = "Task" [ 641.326870] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.338521] env[65121]: DEBUG oslo_vmware.api [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.376542] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106018, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.805136} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.376845] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 641.377121] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 641.377771] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9af6c30-f86e-4ab3-b2f0-5761cf244153 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.389295] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 641.389295] env[65121]: value = "task-5106026" [ 641.389295] env[65121]: _type = "Task" [ 641.389295] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.397729] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106021, 'name': ReconfigVM_Task, 'duration_secs': 0.280022} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.398585] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Reconfigured VM instance instance-00000001 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 641.399700] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befe1705-9c16-4163-8980-2d34cf34e194 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.406243] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106026, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.428677] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 5f6e4f46-0745-42c0-a779-6cffb60e21ca/5f6e4f46-0745-42c0-a779-6cffb60e21ca.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 641.429098] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce0ddd1e-9d0a-4b99-846b-f3c27be9666b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.456611] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 641.456611] env[65121]: value = "task-5106027" [ 641.456611] env[65121]: _type = "Task" [ 641.456611] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.467674] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106027, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.474473] env[65121]: DEBUG nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 641.494760] env[65121]: WARNING neutronclient.v2_0.client [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.495506] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.496185] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.519769] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.535382] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106023, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.563495] env[65121]: WARNING neutronclient.v2_0.client [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.564089] env[65121]: WARNING openstack [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.564434] env[65121]: WARNING openstack [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.572463] env[65121]: DEBUG nova.network.neutron [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Successfully created port: 5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 641.634683] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.642207] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.718137] env[65121]: INFO nova.compute.manager [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Took 29.88 seconds to build instance. [ 641.842312] env[65121]: DEBUG oslo_vmware.api [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319306} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.842625] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 641.842743] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 641.842909] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.843094] env[65121]: INFO nova.compute.manager [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 641.843345] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 641.843536] env[65121]: DEBUG nova.compute.manager [-] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 641.843677] env[65121]: DEBUG nova.network.neutron [-] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 641.843940] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.844490] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.845275] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.899536] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106026, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120128} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.902556] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 641.904233] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f6275f-9631-4ebf-8c1d-51d7e3aabc52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.929066] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 641.930478] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf737e95-a1ea-458a-939d-edf4883352d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.946846] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7ac263-4180-481e-ac2d-a481cb654f78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.956181] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ed2a13-3dc8-459e-b253-a1f0bfc4ebca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.962337] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 641.962337] env[65121]: value = "task-5106028" [ 641.962337] env[65121]: _type = "Task" [ 641.962337] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.008911] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127df426-41bf-4f7f-875b-61b85a629e5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.012420] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106027, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.021444] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.046539] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486e2361-f496-406e-82f3-dd2411cc161f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.053116] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106022, 'name': ReconfigVM_Task, 'duration_secs': 0.732141} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.053371] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 46f89cfb-14aa-483d-aa8a-bc91356272ca/46f89cfb-14aa-483d-aa8a-bc91356272ca.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 642.054853] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-218059bb-7d1b-447b-9a80-c1dfcb503def {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.071318] env[65121]: DEBUG nova.compute.provider_tree [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.072982] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106023, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.080693] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 642.080693] env[65121]: value = "task-5106029" [ 642.080693] env[65121]: _type = "Task" [ 642.080693] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.091568] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106029, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.112490] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.112903] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.220614] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3e967c-63da-4cf1-ae79-8d4866d230c6 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.534s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.266481] env[65121]: WARNING openstack [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.266849] env[65121]: WARNING openstack [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.298571] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.474952] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106027, 'name': ReconfigVM_Task, 'duration_secs': 0.920402} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.475679] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 5f6e4f46-0745-42c0-a779-6cffb60e21ca/5f6e4f46-0745-42c0-a779-6cffb60e21ca.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 642.476039] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance '5f6e4f46-0745-42c0-a779-6cffb60e21ca' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 642.482780] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.514453] env[65121]: DEBUG nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 642.540618] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106023, 'name': CreateSnapshot_Task, 'duration_secs': 1.04963} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.542948] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 642.545450] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4888b1-5d4e-44ea-bfa5-20a3a3f0471a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.551900] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 642.552160] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 642.552328] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 642.552510] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 642.552664] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 642.552818] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 642.553083] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.553308] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 642.553490] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 642.553694] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 642.553943] env[65121]: DEBUG nova.virt.hardware [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 642.555121] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb310b36-b653-468c-8fb5-201ae167e93b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.573090] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3557c72-a788-47b6-be07-426f963e2575 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.579303] env[65121]: DEBUG nova.scheduler.client.report [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 642.603776] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106029, 'name': Rename_Task, 'duration_secs': 0.337291} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.604416] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 642.604721] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98149155-fd5f-4979-9c40-dd8a166e3406 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.613405] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 642.613405] env[65121]: value = "task-5106030" [ 642.613405] env[65121]: _type = "Task" [ 642.613405] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.627191] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106030, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.655359] env[65121]: WARNING neutronclient.v2_0.client [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.656031] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.656405] env[65121]: WARNING openstack [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.697913] env[65121]: WARNING neutronclient.v2_0.client [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.698646] env[65121]: WARNING openstack [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.699031] env[65121]: WARNING openstack [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.723326] env[65121]: DEBUG nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 642.864500] env[65121]: DEBUG nova.network.neutron [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Updated VIF entry in instance network info cache for port 16314fe2-22b5-437b-aebf-58bf042902d0. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 642.864912] env[65121]: DEBUG nova.network.neutron [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Updating instance_info_cache with network_info: [{"id": "16314fe2-22b5-437b-aebf-58bf042902d0", "address": "fa:16:3e:30:56:f3", "network": {"id": "753fce83-9625-407b-ad4a-3a8c260fad09", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-642414393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ca984047abf4272b00b44b551249fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06eaa4c9-dbc2-4d38-a844-7bf76e7b5a64", "external-id": "nsx-vlan-transportzone-804", "segmentation_id": 804, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16314fe2-22", "ovs_interfaceid": "16314fe2-22b5-437b-aebf-58bf042902d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.968194] env[65121]: DEBUG nova.network.neutron [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updated VIF entry in instance network info cache for port 13f8bd13-f8f0-4818-8dd6-845e9644a395. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 642.968563] env[65121]: DEBUG nova.network.neutron [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.979815] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106028, 'name': ReconfigVM_Task, 'duration_secs': 0.667758} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.980159] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4/20396b81-2041-4bf2-85c4-8ea66c838ec4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 642.980800] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43fcbba2-eb03-4eb6-8773-468eb8718ec5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.986745] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-507f164d-784f-412c-9780-dfb6c84d7525 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.993496] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 642.993496] env[65121]: value = "task-5106031" [ 642.993496] env[65121]: _type = "Task" [ 642.993496] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.011848] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985e0ba8-b305-43fb-aab9-96781ae413c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.018046] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106031, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.035216] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance '5f6e4f46-0745-42c0-a779-6cffb60e21ca' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 643.077313] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 643.077313] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-906fc554-9ff9-48de-879f-40b5ee07b653 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.086245] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 643.086348] env[65121]: DEBUG nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 643.092219] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.057s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.093380] env[65121]: INFO nova.compute.claims [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.096015] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 643.096015] env[65121]: value = "task-5106032" [ 643.096015] env[65121]: _type = "Task" [ 643.096015] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.106851] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106032, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.121070] env[65121]: DEBUG nova.network.neutron [-] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 643.129254] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106030, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.250634] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.290169] env[65121]: DEBUG nova.network.neutron [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Successfully updated port: 5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 643.368122] env[65121]: DEBUG oslo_concurrency.lockutils [req-615c5ca2-f1f4-4bed-8c2d-417eab1490b4 req-7f2a5f9c-003a-4520-9da9-da28e21a6431 service nova] Releasing lock "refresh_cache-46f89cfb-14aa-483d-aa8a-bc91356272ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.402410] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.402774] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.475279] env[65121]: DEBUG oslo_concurrency.lockutils [req-d0a1b313-6085-4c5f-9ed4-52ad6f5f267d req-7e3a8543-2d60-4afc-a02d-e904b8a86947 service nova] Releasing lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.508771] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106031, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.544514] env[65121]: WARNING neutronclient.v2_0.client [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.600383] env[65121]: DEBUG nova.compute.utils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 643.610772] env[65121]: DEBUG nova.network.neutron [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Port fc77c1b5-9d9b-479a-a553-e7961313af08 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 643.613594] env[65121]: DEBUG nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 643.615228] env[65121]: DEBUG nova.network.neutron [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 643.615228] env[65121]: WARNING neutronclient.v2_0.client [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.615228] env[65121]: WARNING neutronclient.v2_0.client [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.616199] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.616666] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.637431] env[65121]: INFO nova.compute.manager [-] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Took 1.79 seconds to deallocate network for instance. [ 643.653042] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106032, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.657052] env[65121]: DEBUG oslo_vmware.api [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106030, 'name': PowerOnVM_Task, 'duration_secs': 0.695107} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.657166] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 643.657310] env[65121]: INFO nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Took 9.89 seconds to spawn the instance on the hypervisor. [ 643.657487] env[65121]: DEBUG nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 643.658630] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cf8a97-e29d-4e8d-a499-7828d323a793 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.747761] env[65121]: DEBUG nova.policy [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c865e09a2a0745948db257155fe8a188', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '741e03af2f4d451d9ed77004b644deb8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 643.793480] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.793666] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquired lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 643.793829] env[65121]: DEBUG nova.network.neutron [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 644.007241] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106031, 'name': Rename_Task, 'duration_secs': 0.643894} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.007651] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.008095] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b75be629-681a-4c22-a0bb-c65c61025c41 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.016384] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Waiting for the task: (returnval){ [ 644.016384] env[65121]: value = "task-5106033" [ 644.016384] env[65121]: _type = "Task" [ 644.016384] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.026630] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.118760] env[65121]: DEBUG nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 644.130085] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106032, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.163080] env[65121]: DEBUG nova.network.neutron [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Successfully created port: b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 644.164967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.165965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.165965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.167205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.167765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.168572] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.180967] env[65121]: INFO nova.compute.manager [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Took 30.80 seconds to build instance. [ 644.297344] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.297727] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.351439] env[65121]: DEBUG nova.compute.manager [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Received event network-changed-3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 644.351638] env[65121]: DEBUG nova.compute.manager [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Refreshing instance network info cache due to event network-changed-3a66c1d6-9549-459d-bb1b-66781725a57f. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 644.351902] env[65121]: DEBUG oslo_concurrency.lockutils [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Acquiring lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.352064] env[65121]: DEBUG oslo_concurrency.lockutils [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Acquired lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.352233] env[65121]: DEBUG nova.network.neutron [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Refreshing network info cache for port 3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 644.379010] env[65121]: DEBUG nova.network.neutron [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 644.417744] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.418157] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.514047] env[65121]: WARNING neutronclient.v2_0.client [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.514047] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.514322] env[65121]: WARNING openstack [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.538211] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106033, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.583288] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb6c663-6217-4f5b-b72c-d3538e2995b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.593218] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9221b1-2e03-4f58-a6ee-ff3ea3b3d685 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.636928] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75acb6da-1c0a-460c-bf6a-00951b14079b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.645428] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106032, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.648829] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d30e582-f7f9-47fc-b776-2a8e13bb3064 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.665669] env[65121]: DEBUG nova.compute.provider_tree [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.683173] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d20f9677-ef9e-4bb3-967e-f8e4489a4b06 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.571s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.696603] env[65121]: DEBUG nova.network.neutron [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 644.744793] env[65121]: DEBUG nova.compute.manager [req-0790f92e-d4b6-4b99-b5cc-226c3a727323 req-d532681c-2351-479c-b4b1-fd55f126d1ca service nova] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Received event network-vif-deleted-1542237c-e4ab-42a7-a285-090f397a4a70 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 644.856052] env[65121]: WARNING neutronclient.v2_0.client [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.857619] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.858876] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.032709] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106033, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.138061] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106032, 'name': CloneVM_Task, 'duration_secs': 1.910831} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.138061] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Created linked-clone VM from snapshot [ 645.138622] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3c5e07-ef15-46c4-97ec-cabbbe8c682f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.148313] env[65121]: DEBUG nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 645.156424] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Uploading image e026d22f-eae3-4201-a3e5-3ca70f85ce90 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 645.166902] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.167325] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.175642] env[65121]: WARNING neutronclient.v2_0.client [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.178178] env[65121]: DEBUG nova.scheduler.client.report [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 645.186950] env[65121]: DEBUG nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 645.191651] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:20:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='203556608',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-735823413',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 645.192323] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 645.192323] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 645.192323] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 645.192464] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 645.192510] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 645.192691] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.192841] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 645.192994] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 645.193164] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 645.193405] env[65121]: DEBUG nova.virt.hardware [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 645.195572] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39ab226-1fa7-4f5d-ae5b-cf4829f323c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.199040] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Releasing lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.199348] env[65121]: DEBUG nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Instance network_info: |[{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 645.200366] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:3e:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a9abd00f-2cea-40f8-9804-a56b6431192d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5604925f-4cea-418a-a42d-502bac3a6114', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 645.208885] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Creating folder: Project (0ede261894564e9d9bf50ce9cddeec77). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 645.212295] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 645.212295] env[65121]: value = "vm-993315" [ 645.212295] env[65121]: _type = "VirtualMachine" [ 645.212295] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 645.212944] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae8fffa9-981e-4588-af22-9cc8261410c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.215495] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4d787a82-758d-4a36-82fe-eb1bf501f2f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.221191] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ba9290-6d34-4738-a00a-8122a9e25bd0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.227698] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease: (returnval){ [ 645.227698] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529be562-9665-e430-2bdd-d0830534e8fa" [ 645.227698] env[65121]: _type = "HttpNfcLease" [ 645.227698] env[65121]: } obtained for exporting VM: (result){ [ 645.227698] env[65121]: value = "vm-993315" [ 645.227698] env[65121]: _type = "VirtualMachine" [ 645.227698] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 645.228229] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the lease: (returnval){ [ 645.228229] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529be562-9665-e430-2bdd-d0830534e8fa" [ 645.228229] env[65121]: _type = "HttpNfcLease" [ 645.228229] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 645.242354] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Created folder: Project (0ede261894564e9d9bf50ce9cddeec77) in parent group-v993268. [ 645.242568] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Creating folder: Instances. Parent ref: group-v993316. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 645.243378] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ed212ba-d089-42f7-b69d-49aed5c197ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.246742] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 645.246742] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529be562-9665-e430-2bdd-d0830534e8fa" [ 645.246742] env[65121]: _type = "HttpNfcLease" [ 645.246742] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 645.252018] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.252192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.252363] env[65121]: DEBUG nova.network.neutron [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 645.257758] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Created folder: Instances in parent group-v993316. [ 645.257758] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 645.258408] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 645.258806] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84bfff5a-bac6-4614-a8fb-e91119cb91c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.294411] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 645.294411] env[65121]: value = "task-5106037" [ 645.294411] env[65121]: _type = "Task" [ 645.294411] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.304372] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106037, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.356401] env[65121]: WARNING neutronclient.v2_0.client [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.357339] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.357752] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.533137] env[65121]: DEBUG oslo_vmware.api [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Task: {'id': task-5106033, 'name': PowerOnVM_Task, 'duration_secs': 1.33656} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.533417] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.533699] env[65121]: DEBUG nova.compute.manager [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 645.534479] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52992a27-16d7-49a4-aa21-275f90c37a20 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.685905] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.686533] env[65121]: DEBUG nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 645.690391] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.559s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.690610] env[65121]: DEBUG nova.objects.instance [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 645.722775] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.738231] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 645.738231] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529be562-9665-e430-2bdd-d0830534e8fa" [ 645.738231] env[65121]: _type = "HttpNfcLease" [ 645.738231] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 645.742477] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 645.742477] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529be562-9665-e430-2bdd-d0830534e8fa" [ 645.742477] env[65121]: _type = "HttpNfcLease" [ 645.742477] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 645.742477] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e369b1-a3b8-45fc-bfaa-043727ffec9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.745819] env[65121]: DEBUG nova.network.neutron [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Successfully updated port: b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 645.753276] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5215f020-f4ce-aa3a-b94c-61267edd0178/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 645.753380] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5215f020-f4ce-aa3a-b94c-61267edd0178/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 645.755762] env[65121]: WARNING neutronclient.v2_0.client [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.755997] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.756680] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.767949] env[65121]: DEBUG nova.network.neutron [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updated VIF entry in instance network info cache for port 3a66c1d6-9549-459d-bb1b-66781725a57f. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 645.768298] env[65121]: DEBUG nova.network.neutron [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updating instance_info_cache with network_info: [{"id": "3a66c1d6-9549-459d-bb1b-66781725a57f", "address": "fa:16:3e:be:bf:37", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66c1d6-95", "ovs_interfaceid": "3a66c1d6-9549-459d-bb1b-66781725a57f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 645.839811] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106037, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.871294] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dc945afb-d5b2-4bf4-83f9-fad353812e01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.996582] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.996968] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.050327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.174395] env[65121]: WARNING neutronclient.v2_0.client [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.175148] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.175486] env[65121]: WARNING openstack [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.205568] env[65121]: DEBUG nova.compute.utils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 646.209133] env[65121]: DEBUG nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 646.209444] env[65121]: DEBUG nova.network.neutron [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 646.209781] env[65121]: WARNING neutronclient.v2_0.client [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.210087] env[65121]: WARNING neutronclient.v2_0.client [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.210775] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.211284] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.221475] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fef58b6a-1fcb-4bc4-825e-f234c3a8058b tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.530s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.221475] env[65121]: DEBUG nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 646.223905] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.976s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.225733] env[65121]: INFO nova.compute.claims [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.253932] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.255108] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.255664] env[65121]: DEBUG nova.network.neutron [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 646.271912] env[65121]: DEBUG oslo_concurrency.lockutils [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Releasing lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.272287] env[65121]: DEBUG nova.compute.manager [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Received event network-changed-3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 646.272481] env[65121]: DEBUG nova.compute.manager [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Refreshing instance network info cache due to event network-changed-3a66c1d6-9549-459d-bb1b-66781725a57f. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 646.273016] env[65121]: DEBUG oslo_concurrency.lockutils [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Acquiring lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.273016] env[65121]: DEBUG oslo_concurrency.lockutils [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Acquired lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.273152] env[65121]: DEBUG nova.network.neutron [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Refreshing network info cache for port 3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 646.349793] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106037, 'name': CreateVM_Task, 'duration_secs': 0.610408} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.353124] env[65121]: DEBUG nova.network.neutron [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance_info_cache with network_info: [{"id": "fc77c1b5-9d9b-479a-a553-e7961313af08", "address": "fa:16:3e:f6:e5:10", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.52", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc77c1b5-9d", "ovs_interfaceid": "fc77c1b5-9d9b-479a-a553-e7961313af08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.357134] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 646.361090] env[65121]: WARNING neutronclient.v2_0.client [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.361604] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.361734] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.362125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 646.362624] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15eca472-6f87-4467-a407-3f7dea96f558 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.371271] env[65121]: DEBUG nova.policy [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '390615f3a68d489183e81f8c888e5247', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '018269ed0c644392aebdb56cf0523292', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 646.376300] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 646.376300] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5210c53e-f316-b5bc-497d-02f7a68f3142" [ 646.376300] env[65121]: _type = "Task" [ 646.376300] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.388177] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5210c53e-f316-b5bc-497d-02f7a68f3142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.717909] env[65121]: DEBUG nova.network.neutron [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Successfully created port: 18509dce-9df7-41cc-af3e-70f419fbb5d5 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 646.760890] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.761856] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.776672] env[65121]: WARNING neutronclient.v2_0.client [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.777662] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.778041] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.861966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.888163] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5210c53e-f316-b5bc-497d-02f7a68f3142, 'name': SearchDatastore_Task, 'duration_secs': 0.01858} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.888847] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.889240] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 646.889722] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.889992] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.890329] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 646.890913] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da7015e9-4b55-40e7-baab-d294f295a9be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.905197] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 646.906161] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 646.907513] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-687cb933-1ee7-49eb-b7e9-11c2e09404a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.916374] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 646.916374] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f0ae7c-019f-23f6-8a6e-6ed2967eb744" [ 646.916374] env[65121]: _type = "Task" [ 646.916374] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.930585] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f0ae7c-019f-23f6-8a6e-6ed2967eb744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.241535] env[65121]: DEBUG nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 647.335332] env[65121]: DEBUG nova.network.neutron [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 647.396181] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7717fa-7263-4e8b-aa54-ea7111f92832 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.424369] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18367ebd-05c2-43e4-9e33-1d4817d36772 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.438450] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f0ae7c-019f-23f6-8a6e-6ed2967eb744, 'name': SearchDatastore_Task, 'duration_secs': 0.014492} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.439069] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance '5f6e4f46-0745-42c0-a779-6cffb60e21ca' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 647.446714] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7b01e12-168c-46c3-998b-ee9a2fae476f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.452850] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 647.452850] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c7cc65-457f-3e37-baf0-07b331004bd7" [ 647.452850] env[65121]: _type = "Task" [ 647.452850] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.463213] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c7cc65-457f-3e37-baf0-07b331004bd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.675226] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6d8452-9c0a-454e-b34d-31cfaa681eee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.688666] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048018c5-d49e-4ccb-a549-29ff1fd580f2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.746077] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe32b9c8-8a62-460f-97b1-cdefc0d1c1b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.761967] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078cf52e-5c81-4348-802a-362af723819f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.769149] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 647.769508] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 647.769508] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 647.769627] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 647.770587] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 647.770587] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 647.770587] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.770587] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 647.770864] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 647.771086] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 647.771397] env[65121]: DEBUG nova.virt.hardware [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 647.772723] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a252703-15fe-41c9-8278-4f4cdea37a7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.796385] env[65121]: DEBUG nova.compute.provider_tree [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.800059] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b36b0b-09fc-4a6e-865b-a073b97060fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.824930] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.825703] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.946358] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 647.946795] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f9011ab-ea2b-4724-aeee-13bea0bb6580 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.960021] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 647.960021] env[65121]: value = "task-5106038" [ 647.960021] env[65121]: _type = "Task" [ 647.960021] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.969495] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c7cc65-457f-3e37-baf0-07b331004bd7, 'name': SearchDatastore_Task, 'duration_secs': 0.030306} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.970618] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.971168] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 5854d21a-d1a8-4043-aec8-b37ff25c40e7/5854d21a-d1a8-4043-aec8-b37ff25c40e7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 647.971909] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f2c8022-524f-4a2e-8ba5-56e744d8b53c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.979875] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106038, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.989022] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 647.989022] env[65121]: value = "task-5106039" [ 647.989022] env[65121]: _type = "Task" [ 647.989022] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.000816] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.308602] env[65121]: DEBUG nova.scheduler.client.report [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 648.367211] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.371167] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.475781] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106038, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.502352] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106039, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.507898] env[65121]: DEBUG nova.network.neutron [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Successfully updated port: 18509dce-9df7-41cc-af3e-70f419fbb5d5 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 648.534553] env[65121]: WARNING neutronclient.v2_0.client [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.535163] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.535654] env[65121]: WARNING openstack [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.814853] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.817014] env[65121]: DEBUG nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 648.818452] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.238s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.820023] env[65121]: INFO nova.compute.claims [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.828683] env[65121]: DEBUG nova.compute.manager [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 648.829102] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b86ee6-9194-4fa3-84cd-c23fc0698fd3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.840517] env[65121]: WARNING neutronclient.v2_0.client [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.841152] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.841475] env[65121]: WARNING openstack [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.974829] env[65121]: DEBUG oslo_vmware.api [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106038, 'name': PowerOnVM_Task, 'duration_secs': 0.746138} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.975601] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 648.975601] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f923096c-749f-4b77-8f90-e1bf264304ca tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance '5f6e4f46-0745-42c0-a779-6cffb60e21ca' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 649.003866] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106039, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.717834} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.006640] env[65121]: DEBUG nova.network.neutron [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Updating instance_info_cache with network_info: [{"id": "b25aa25b-2d1b-414e-883a-324894d4b483", "address": "fa:16:3e:0e:be:f6", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25aa25b-2d", "ovs_interfaceid": "b25aa25b-2d1b-414e-883a-324894d4b483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 649.008318] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 5854d21a-d1a8-4043-aec8-b37ff25c40e7/5854d21a-d1a8-4043-aec8-b37ff25c40e7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 649.008467] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 649.012174] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8a5cd46-f991-4a0b-84ff-2dbdb6d0951b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.014970] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "refresh_cache-6a4841b6-6076-4e9c-a6cf-a658957e2931" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.015129] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired lock "refresh_cache-6a4841b6-6076-4e9c-a6cf-a658957e2931" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 649.015303] env[65121]: DEBUG nova.network.neutron [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 649.024407] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 649.024407] env[65121]: value = "task-5106040" [ 649.024407] env[65121]: _type = "Task" [ 649.024407] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.038543] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106040, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.214224] env[65121]: DEBUG nova.network.neutron [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updated VIF entry in instance network info cache for port 3a66c1d6-9549-459d-bb1b-66781725a57f. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 649.214853] env[65121]: DEBUG nova.network.neutron [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updating instance_info_cache with network_info: [{"id": "3a66c1d6-9549-459d-bb1b-66781725a57f", "address": "fa:16:3e:be:bf:37", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66c1d6-95", "ovs_interfaceid": "3a66c1d6-9549-459d-bb1b-66781725a57f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 649.325050] env[65121]: DEBUG nova.compute.utils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 649.330758] env[65121]: DEBUG nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 649.331210] env[65121]: DEBUG nova.network.neutron [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 649.332342] env[65121]: WARNING neutronclient.v2_0.client [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.333167] env[65121]: WARNING neutronclient.v2_0.client [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.335267] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.335267] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.350687] env[65121]: INFO nova.compute.manager [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] instance snapshotting [ 649.355042] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e5277f-52ec-43bc-8530-0b06d94a66c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.380065] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1943bcf-d49b-4cfe-9ab6-5992869e64be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.427206] env[65121]: DEBUG nova.policy [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bf232407a7346c98ea64ed673a179f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b35af4e2b6844a9081f34466661c83b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 649.510147] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Releasing lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 649.510634] env[65121]: DEBUG nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Instance network_info: |[{"id": "b25aa25b-2d1b-414e-883a-324894d4b483", "address": "fa:16:3e:0e:be:f6", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25aa25b-2d", "ovs_interfaceid": "b25aa25b-2d1b-414e-883a-324894d4b483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 649.511235] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:be:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b25aa25b-2d1b-414e-883a-324894d4b483', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 649.523061] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Creating folder: Project (741e03af2f4d451d9ed77004b644deb8). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 649.524654] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.525042] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.532807] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fbf4277-1aac-4d36-af73-562d8b9508fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.547079] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106040, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.141604} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.547947] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 649.549126] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f25af09-3e76-425a-a5f6-037e7cb29226 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.553166] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Created folder: Project (741e03af2f4d451d9ed77004b644deb8) in parent group-v993268. [ 649.553490] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Creating folder: Instances. Parent ref: group-v993319. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 649.554115] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-222dfb04-9a1c-4eb6-b0b0-64dd72870384 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.577175] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] 5854d21a-d1a8-4043-aec8-b37ff25c40e7/5854d21a-d1a8-4043-aec8-b37ff25c40e7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 649.579877] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57583c82-8d2e-484e-9378-9d54616b692d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.596378] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Created folder: Instances in parent group-v993319. [ 649.596676] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 649.596901] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 649.597610] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80849b26-7795-4857-a7d6-7e7d80c40442 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.619419] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 649.619419] env[65121]: value = "task-5106043" [ 649.619419] env[65121]: _type = "Task" [ 649.619419] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.625864] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 649.625864] env[65121]: value = "task-5106044" [ 649.625864] env[65121]: _type = "Task" [ 649.625864] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.632247] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106043, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.638483] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106044, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.718111] env[65121]: DEBUG oslo_concurrency.lockutils [req-4fb67d53-c21e-46c2-ab28-d751c6b1b24f req-ae75a60a-34c9-406b-8186-624211954393 service nova] Releasing lock "refresh_cache-f1a5fdce-a940-43d4-83d7-716786c9ac34" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 649.831345] env[65121]: DEBUG nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 649.894081] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 649.895206] env[65121]: DEBUG nova.network.neutron [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Successfully created port: 33c6c3ae-feb6-4512-ac82-46c27afad318 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 649.897651] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a8dfff81-a63a-4f44-aeac-8aa265bb8433 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.907399] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 649.907399] env[65121]: value = "task-5106045" [ 649.907399] env[65121]: _type = "Task" [ 649.907399] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.921879] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106045, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.978231] env[65121]: DEBUG nova.network.neutron [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 650.136913] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106043, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.143329] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106044, 'name': CreateVM_Task, 'duration_secs': 0.511503} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.143940] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 650.144516] env[65121]: WARNING neutronclient.v2_0.client [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.144911] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.145091] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.145551] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 650.146275] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cbcdf2e-ec99-45c0-bab8-8470783c1b53 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.154108] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 650.154108] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b35faa-8c85-de99-e5f7-13057b443ec7" [ 650.154108] env[65121]: _type = "Task" [ 650.154108] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.163609] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b35faa-8c85-de99-e5f7-13057b443ec7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.306656] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0f247a-8b22-4d23-a78a-172dc4df4cc7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.315622] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a935800b-1afd-4259-a4b8-578ecbecb97d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.356937] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca2628d-2f0f-4889-89af-fb192a7be3a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.367022] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892357ff-dadd-4f1f-9724-881e645ba96e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.379157] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.379157] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.401426] env[65121]: DEBUG nova.compute.provider_tree [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.403174] env[65121]: DEBUG nova.compute.manager [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Received event network-vif-deleted-062568cf-fd2f-407b-8cf8-da76b66e5c30 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 650.403400] env[65121]: DEBUG nova.compute.manager [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Received event network-changed-4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 650.403553] env[65121]: DEBUG nova.compute.manager [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Refreshing instance network info cache due to event network-changed-4ea2aa90-9370-491b-97b5-2c112a535219. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 650.403806] env[65121]: DEBUG oslo_concurrency.lockutils [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.405666] env[65121]: DEBUG oslo_concurrency.lockutils [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.405666] env[65121]: DEBUG nova.network.neutron [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Refreshing network info cache for port 4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 650.422785] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106045, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.636036] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106043, 'name': ReconfigVM_Task, 'duration_secs': 0.53441} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.636368] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Reconfigured VM instance instance-00000010 to attach disk [datastore1] 5854d21a-d1a8-4043-aec8-b37ff25c40e7/5854d21a-d1a8-4043-aec8-b37ff25c40e7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 650.637571] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94a06d14-3bcd-4cf7-a9c5-955173f76e48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.647043] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 650.647043] env[65121]: value = "task-5106046" [ 650.647043] env[65121]: _type = "Task" [ 650.647043] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.659320] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106046, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.672820] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b35faa-8c85-de99-e5f7-13057b443ec7, 'name': SearchDatastore_Task, 'duration_secs': 0.017062} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.673268] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.673591] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 650.673934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.674141] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.674713] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 650.674713] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0717471f-9977-4933-97fa-8049b4b2138e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.687141] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 650.687425] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 650.688447] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99861dea-db70-4f8c-9bf0-5aed0a1b7fed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.697018] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 650.697018] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f108e9-7aec-1421-e870-9783dfcac347" [ 650.697018] env[65121]: _type = "Task" [ 650.697018] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.707978] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f108e9-7aec-1421-e870-9783dfcac347, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.860372] env[65121]: DEBUG nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 650.889944] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 650.890257] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 650.890507] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 650.890770] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 650.890952] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 650.891165] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 650.891476] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.891731] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 650.891966] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 650.892197] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 650.892483] env[65121]: DEBUG nova.virt.hardware [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 650.893727] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2af3386-4fd9-4b06-b297-ed57e89d2318 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.903457] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07381e6d-524d-48cb-87dc-6d3c079297e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.910575] env[65121]: DEBUG nova.scheduler.client.report [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 650.914313] env[65121]: WARNING neutronclient.v2_0.client [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.914923] env[65121]: WARNING openstack [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.915347] env[65121]: WARNING openstack [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.934237] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106045, 'name': CreateSnapshot_Task, 'duration_secs': 1.003554} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.946446] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 650.948067] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bfbec3-9cbb-47b7-bb58-366c9b4a3bdf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.158414] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106046, 'name': Rename_Task, 'duration_secs': 0.235544} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.158690] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 651.158941] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e43c7b78-346e-438f-b1e9-724599e15188 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.167612] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 651.167612] env[65121]: value = "task-5106047" [ 651.167612] env[65121]: _type = "Task" [ 651.167612] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.177713] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106047, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.208639] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f108e9-7aec-1421-e870-9783dfcac347, 'name': SearchDatastore_Task, 'duration_secs': 0.012769} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.209371] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-964ca073-cae0-4b03-8e8e-caece5bda6e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.215966] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 651.215966] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a24fbd-7640-b609-8540-7aadc4c833b3" [ 651.215966] env[65121]: _type = "Task" [ 651.215966] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.226855] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a24fbd-7640-b609-8540-7aadc4c833b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.426506] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 651.426506] env[65121]: DEBUG nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 651.428780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.059s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.431557] env[65121]: INFO nova.compute.claims [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.471129] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 651.472706] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f51a7891-2147-4090-a0a6-82d740961bb2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.488023] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 651.488023] env[65121]: value = "task-5106048" [ 651.488023] env[65121]: _type = "Task" [ 651.488023] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.497769] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106048, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.569102] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Received event network-vif-plugged-5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 651.569165] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Acquiring lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.569344] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.569470] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 651.569728] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] No waiting events found dispatching network-vif-plugged-5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 651.569765] env[65121]: WARNING nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Received unexpected event network-vif-plugged-5604925f-4cea-418a-a42d-502bac3a6114 for instance with vm_state building and task_state spawning. [ 651.570304] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Received event network-changed-5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 651.570738] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing instance network info cache due to event network-changed-5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 651.571390] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Acquiring lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.571570] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Acquired lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.571744] env[65121]: DEBUG nova.network.neutron [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 651.681874] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106047, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.708329] env[65121]: DEBUG nova.network.neutron [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Successfully updated port: 33c6c3ae-feb6-4512-ac82-46c27afad318 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 651.731329] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a24fbd-7640-b609-8540-7aadc4c833b3, 'name': SearchDatastore_Task, 'duration_secs': 0.01981} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.732890] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.733238] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e2d32a5e-c350-4b2b-9243-c3b412193a82/e2d32a5e-c350-4b2b-9243-c3b412193a82.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 651.733768] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a8dbdd3-527b-44ab-94d0-8045623b0bca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.742881] env[65121]: WARNING neutronclient.v2_0.client [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.743579] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.744020] env[65121]: WARNING openstack [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.758635] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 651.758635] env[65121]: value = "task-5106049" [ 651.758635] env[65121]: _type = "Task" [ 651.758635] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.771822] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106049, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.944462] env[65121]: DEBUG nova.compute.utils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 651.947392] env[65121]: DEBUG nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 651.948253] env[65121]: DEBUG nova.network.neutron [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 651.948424] env[65121]: WARNING neutronclient.v2_0.client [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.949061] env[65121]: WARNING neutronclient.v2_0.client [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.949953] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.950470] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.001770] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106048, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.026945] env[65121]: WARNING openstack [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.027388] env[65121]: WARNING openstack [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.065767] env[65121]: DEBUG nova.network.neutron [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Updating instance_info_cache with network_info: [{"id": "18509dce-9df7-41cc-af3e-70f419fbb5d5", "address": "fa:16:3e:3e:13:c2", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.95", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18509dce-9d", "ovs_interfaceid": "18509dce-9df7-41cc-af3e-70f419fbb5d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 652.077806] env[65121]: WARNING neutronclient.v2_0.client [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.078691] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.079295] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.147637] env[65121]: DEBUG nova.policy [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e811ab6f70dc4249aff2d47a11f3baf8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b816d77d6f3b4bc7a20d5b97ba068006', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 652.190557] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106047, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.214155] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "refresh_cache-7a0d0769-31c5-4ae1-8520-a744f64d39c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.214155] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "refresh_cache-7a0d0769-31c5-4ae1-8520-a744f64d39c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.214155] env[65121]: DEBUG nova.network.neutron [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 652.273552] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106049, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.461615] env[65121]: DEBUG nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 652.503934] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106048, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.510555] env[65121]: WARNING neutronclient.v2_0.client [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 652.510890] env[65121]: WARNING openstack [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.511260] env[65121]: WARNING openstack [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.571173] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Releasing lock "refresh_cache-6a4841b6-6076-4e9c-a6cf-a658957e2931" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.571173] env[65121]: DEBUG nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Instance network_info: |[{"id": "18509dce-9df7-41cc-af3e-70f419fbb5d5", "address": "fa:16:3e:3e:13:c2", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.95", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18509dce-9d", "ovs_interfaceid": "18509dce-9df7-41cc-af3e-70f419fbb5d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 652.571383] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:13:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18509dce-9df7-41cc-af3e-70f419fbb5d5', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 652.580354] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 652.582586] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 652.582967] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0b6b0c1-8cf6-440a-9a35-7c15a29d8358 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.615254] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 652.615254] env[65121]: value = "task-5106050" [ 652.615254] env[65121]: _type = "Task" [ 652.615254] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.625957] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106050, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.683425] env[65121]: DEBUG oslo_vmware.api [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106047, 'name': PowerOnVM_Task, 'duration_secs': 1.323262} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.686530] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 652.687517] env[65121]: INFO nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Took 10.17 seconds to spawn the instance on the hypervisor. [ 652.687517] env[65121]: DEBUG nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 652.688790] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51270ff0-6074-4262-8f0b-9e08ca1cc556 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.692418] env[65121]: DEBUG nova.network.neutron [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Successfully created port: a27cb922-5612-41b7-9637-f2c0bff9c8b5 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 652.723345] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.723855] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.776444] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106049, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647388} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.776728] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e2d32a5e-c350-4b2b-9243-c3b412193a82/e2d32a5e-c350-4b2b-9243-c3b412193a82.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 652.776946] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 652.778503] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb9a88be-494d-4b45-9e40-4a0687a69847 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.789386] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 652.789386] env[65121]: value = "task-5106051" [ 652.789386] env[65121]: _type = "Task" [ 652.789386] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.802379] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106051, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.907612] env[65121]: DEBUG nova.compute.manager [None req-a441d32d-2253-436d-9b21-a43ce8222892 tempest-ServerDiagnosticsV248Test-217477461 tempest-ServerDiagnosticsV248Test-217477461-project-admin] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 652.912891] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e2160d-22ca-4528-a6a9-af437811cc3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.924928] env[65121]: INFO nova.compute.manager [None req-a441d32d-2253-436d-9b21-a43ce8222892 tempest-ServerDiagnosticsV248Test-217477461 tempest-ServerDiagnosticsV248Test-217477461-project-admin] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Retrieving diagnostics [ 652.926473] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924d4b7f-ef48-4628-9c48-e5b6e147b5cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.009976] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106048, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.057608] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "20396b81-2041-4bf2-85c4-8ea66c838ec4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.057912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "20396b81-2041-4bf2-85c4-8ea66c838ec4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.058143] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "20396b81-2041-4bf2-85c4-8ea66c838ec4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.058343] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "20396b81-2041-4bf2-85c4-8ea66c838ec4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.058505] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "20396b81-2041-4bf2-85c4-8ea66c838ec4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.061587] env[65121]: INFO nova.compute.manager [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Terminating instance [ 653.113845] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da188245-d6cf-426f-824e-8f1cd67996df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.133301] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106050, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.134701] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d18fa0-7ad3-40a0-ad05-0cf24830fb97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.177199] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2767b0c7-d868-463b-be83-d800d77a4a23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.190420] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611e50ea-c4af-4d77-913d-6f1fda6bbbf8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.210290] env[65121]: DEBUG nova.compute.provider_tree [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.218812] env[65121]: INFO nova.compute.manager [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Took 28.84 seconds to build instance. [ 653.228483] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.228910] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.305191] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106051, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13717} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.305191] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 653.305191] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344a5315-b09c-4ee7-9110-d9e07815a4d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.334480] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] e2d32a5e-c350-4b2b-9243-c3b412193a82/e2d32a5e-c350-4b2b-9243-c3b412193a82.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 653.335258] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-441404cf-ffef-4224-85b6-af09eca81612 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.357129] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 653.357129] env[65121]: value = "task-5106052" [ 653.357129] env[65121]: _type = "Task" [ 653.357129] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.367843] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.481249] env[65121]: DEBUG nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 653.510804] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106048, 'name': CloneVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.524931] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 653.525129] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 653.525885] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 653.525885] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 653.525885] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 653.525885] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 653.526344] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.526537] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 653.526718] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 653.527650] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 653.527702] env[65121]: DEBUG nova.virt.hardware [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 653.528719] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbac0eb9-82d4-4159-8d8a-01ae8ff69ce5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.538349] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99adbf30-0457-4c87-b64f-f48a5f8c0bb0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.565441] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "refresh_cache-20396b81-2041-4bf2-85c4-8ea66c838ec4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.565649] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquired lock "refresh_cache-20396b81-2041-4bf2-85c4-8ea66c838ec4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 653.565832] env[65121]: DEBUG nova.network.neutron [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 653.629328] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106050, 'name': CreateVM_Task, 'duration_secs': 0.566444} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.629328] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 653.632382] env[65121]: WARNING neutronclient.v2_0.client [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 653.632382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.632382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 653.632382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 653.632382] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e933a1d-cc38-4906-b33a-44dd5b7523aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.637787] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 653.637787] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a37c78-7c2f-430d-da3a-495f1b18f05b" [ 653.637787] env[65121]: _type = "Task" [ 653.637787] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.648074] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a37c78-7c2f-430d-da3a-495f1b18f05b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.720195] env[65121]: DEBUG nova.scheduler.client.report [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 653.725060] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8748f75-4fb9-4681-865d-c3129fd56b74 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.865s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.737973] env[65121]: DEBUG nova.network.neutron [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 653.869285] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106052, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.922256] env[65121]: DEBUG nova.network.neutron [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updated VIF entry in instance network info cache for port 4ea2aa90-9370-491b-97b5-2c112a535219. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 653.922807] env[65121]: DEBUG nova.network.neutron [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 654.008630] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106048, 'name': CloneVM_Task, 'duration_secs': 2.046163} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.009130] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Created linked-clone VM from snapshot [ 654.010647] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee2978d-feeb-4057-9bc2-747c6d4f97b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.019114] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Uploading image 8af4d2e1-314a-42f6-b6eb-dfa13df15f20 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 654.048588] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 654.048588] env[65121]: value = "vm-993323" [ 654.048588] env[65121]: _type = "VirtualMachine" [ 654.048588] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 654.048889] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-846c1972-1e27-46a3-b816-195a0387002a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.058025] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lease: (returnval){ [ 654.058025] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5213383e-3295-e102-cee2-87f1804728c7" [ 654.058025] env[65121]: _type = "HttpNfcLease" [ 654.058025] env[65121]: } obtained for exporting VM: (result){ [ 654.058025] env[65121]: value = "vm-993323" [ 654.058025] env[65121]: _type = "VirtualMachine" [ 654.058025] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 654.058443] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the lease: (returnval){ [ 654.058443] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5213383e-3295-e102-cee2-87f1804728c7" [ 654.058443] env[65121]: _type = "HttpNfcLease" [ 654.058443] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 654.071263] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 654.071263] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5213383e-3295-e102-cee2-87f1804728c7" [ 654.071263] env[65121]: _type = "HttpNfcLease" [ 654.071263] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 654.073321] env[65121]: WARNING neutronclient.v2_0.client [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.073818] env[65121]: WARNING openstack [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.075427] env[65121]: WARNING openstack [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.087837] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.088263] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.148452] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a37c78-7c2f-430d-da3a-495f1b18f05b, 'name': SearchDatastore_Task, 'duration_secs': 0.028607} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.148754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 654.148982] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 654.149228] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.149365] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.149536] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 654.149915] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b5257a1-e245-4713-ad34-f93144d324e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.164113] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 654.165372] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 654.166144] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c980815-3795-4b98-886c-cce3350b32bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.173806] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 654.173806] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52798d7a-912d-6a7b-5f63-c45f05ccd8b2" [ 654.173806] env[65121]: _type = "Task" [ 654.173806] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.183407] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52798d7a-912d-6a7b-5f63-c45f05ccd8b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.223796] env[65121]: WARNING neutronclient.v2_0.client [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.224675] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.229020] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.236988] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.808s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 654.240884] env[65121]: DEBUG nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 654.244540] env[65121]: DEBUG nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 654.246602] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.503s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.249055] env[65121]: INFO nova.compute.claims [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.281469] env[65121]: DEBUG nova.network.neutron [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 654.371509] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106052, 'name': ReconfigVM_Task, 'duration_secs': 0.551889} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.371980] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Reconfigured VM instance instance-00000011 to attach disk [datastore2] e2d32a5e-c350-4b2b-9243-c3b412193a82/e2d32a5e-c350-4b2b-9243-c3b412193a82.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 654.372955] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1b187c38-44d0-46f2-bb49-617162bfa19c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.384205] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 654.384205] env[65121]: value = "task-5106054" [ 654.384205] env[65121]: _type = "Task" [ 654.384205] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.396898] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106054, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.426172] env[65121]: DEBUG oslo_concurrency.lockutils [req-4ef7ea5f-f20a-4a30-8a47-ef1e1506f29e req-db65151b-1be7-4004-a1a8-be91416e6af1 service nova] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 654.567777] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 654.567777] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5213383e-3295-e102-cee2-87f1804728c7" [ 654.567777] env[65121]: _type = "HttpNfcLease" [ 654.567777] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 654.568205] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 654.568205] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5213383e-3295-e102-cee2-87f1804728c7" [ 654.568205] env[65121]: _type = "HttpNfcLease" [ 654.568205] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 654.568810] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870a03e5-c8b0-49a9-baf4-4a812712bd87 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.349197] env[65121]: DEBUG nova.compute.utils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 655.355069] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cde36-1906-eedc-7bdb-0641999b5da4/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 655.355430] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cde36-1906-eedc-7bdb-0641999b5da4/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 655.363504] env[65121]: DEBUG nova.network.neutron [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Successfully updated port: a27cb922-5612-41b7-9637-f2c0bff9c8b5 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 655.367617] env[65121]: WARNING neutronclient.v2_0.client [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.368249] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.368625] env[65121]: WARNING openstack [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.380038] env[65121]: DEBUG nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 655.380324] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 655.380800] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.380800] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.381545] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.381636] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.393885] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "refresh_cache-6218957b-6329-4004-97ca-07231b55937a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.394083] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquired lock "refresh_cache-6218957b-6329-4004-97ca-07231b55937a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.394249] env[65121]: DEBUG nova.network.neutron [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 655.455725] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.459368] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52798d7a-912d-6a7b-5f63-c45f05ccd8b2, 'name': SearchDatastore_Task, 'duration_secs': 0.01107} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.459600] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106054, 'name': Rename_Task, 'duration_secs': 0.349708} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.461717] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 655.462606] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f598d1-9f55-4888-b753-172bcc77fcc5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.465248] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd85fbe7-86c8-416b-b40d-68cccf377f29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.472023] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 655.472023] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bb4d73-5818-ffe0-136e-608fbd65e507" [ 655.472023] env[65121]: _type = "Task" [ 655.472023] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.477074] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 655.477074] env[65121]: value = "task-5106055" [ 655.477074] env[65121]: _type = "Task" [ 655.477074] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.483900] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bb4d73-5818-ffe0-136e-608fbd65e507, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.490593] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106055, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.499192] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e67cb3e6-b4f6-49d9-81e2-b3c7f2251409 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.771296] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5215f020-f4ce-aa3a-b94c-61267edd0178/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 655.774323] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24ef5bb-bab8-4514-a9c5-a375e1b5f80a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.787576] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5215f020-f4ce-aa3a-b94c-61267edd0178/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 655.787970] env[65121]: ERROR oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5215f020-f4ce-aa3a-b94c-61267edd0178/disk-0.vmdk due to incomplete transfer. [ 655.788659] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-73f95505-70e4-47c1-8f70-6e94315c4b07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.799093] env[65121]: DEBUG oslo_vmware.rw_handles [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5215f020-f4ce-aa3a-b94c-61267edd0178/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 655.799315] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Uploaded image e026d22f-eae3-4201-a3e5-3ca70f85ce90 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 655.801694] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 655.802177] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9d7b4495-e202-46c5-9e82-f913d84030fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.810547] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 655.810547] env[65121]: value = "task-5106056" [ 655.810547] env[65121]: _type = "Task" [ 655.810547] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.821040] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106056, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.868027] env[65121]: DEBUG nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 655.898994] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.899684] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.998521] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106055, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.998975] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bb4d73-5818-ffe0-136e-608fbd65e507, 'name': SearchDatastore_Task, 'duration_secs': 0.013804} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.999485] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.999885] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 6a4841b6-6076-4e9c-a6cf-a658957e2931/6a4841b6-6076-4e9c-a6cf-a658957e2931.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 656.000213] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4958151-44b2-43f5-82bd-d84174e2f061 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.010345] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 656.010345] env[65121]: value = "task-5106057" [ 656.010345] env[65121]: _type = "Task" [ 656.010345] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.029859] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.158527] env[65121]: DEBUG nova.network.neutron [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.218478] env[65121]: DEBUG nova.policy [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e8520cf91b64499b55cebfca04cec5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc44d2096bff489193193973fe3a3550', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 656.257646] env[65121]: DEBUG nova.network.neutron [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 656.269229] env[65121]: DEBUG nova.network.neutron [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updated VIF entry in instance network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 656.271188] env[65121]: DEBUG nova.network.neutron [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.330771] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106056, 'name': Destroy_Task, 'duration_secs': 0.435802} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.335200] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Destroyed the VM [ 656.335200] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 656.335733] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-55a7a01e-0133-4562-ae1e-067159026ab5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.343907] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 656.343907] env[65121]: value = "task-5106058" [ 656.343907] env[65121]: _type = "Task" [ 656.343907] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.356597] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106058, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.370256] env[65121]: DEBUG nova.network.neutron [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Updating instance_info_cache with network_info: [{"id": "33c6c3ae-feb6-4512-ac82-46c27afad318", "address": "fa:16:3e:20:ab:a5", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33c6c3ae-fe", "ovs_interfaceid": "33c6c3ae-feb6-4512-ac82-46c27afad318", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.457948] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b36ed7-7e69-4d84-bcc6-25757d3dcd07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.468774] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99c1ab9-8f55-4eea-824d-200c56f4fc29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.513068] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed06a74-a428-476d-aa0e-c86c10e80a33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.525452] env[65121]: DEBUG oslo_vmware.api [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106055, 'name': PowerOnVM_Task, 'duration_secs': 0.769755} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.525843] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 656.526068] env[65121]: INFO nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Took 11.38 seconds to spawn the instance on the hypervisor. [ 656.526263] env[65121]: DEBUG nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 656.527548] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d799b945-16ac-470f-87c2-97d866a66b35 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.535763] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3893eed-323d-4a1c-bd23-51008b5ca928 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.538781] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106057, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.557744] env[65121]: DEBUG nova.compute.provider_tree [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.664193] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Releasing lock "refresh_cache-20396b81-2041-4bf2-85c4-8ea66c838ec4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.664193] env[65121]: DEBUG nova.compute.manager [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 656.664300] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 656.665167] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9761940-74aa-427f-8a25-4b17a3adb788 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.676726] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 656.677330] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26e63db7-8609-4c69-bfc1-70f7d0c5c0b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.685494] env[65121]: DEBUG oslo_vmware.api [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 656.685494] env[65121]: value = "task-5106059" [ 656.685494] env[65121]: _type = "Task" [ 656.685494] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.696207] env[65121]: DEBUG oslo_vmware.api [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5106059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.777489] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Releasing lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.777489] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Received event network-vif-plugged-b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 656.777489] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Acquiring lock "e2d32a5e-c350-4b2b-9243-c3b412193a82-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.777489] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.777966] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.777966] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] No waiting events found dispatching network-vif-plugged-b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 656.778056] env[65121]: WARNING nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Received unexpected event network-vif-plugged-b25aa25b-2d1b-414e-883a-324894d4b483 for instance with vm_state building and task_state spawning. [ 656.778403] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Received event network-changed-b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 656.779067] env[65121]: DEBUG nova.compute.manager [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Refreshing instance network info cache due to event network-changed-b25aa25b-2d1b-414e-883a-324894d4b483. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 656.779067] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Acquiring lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.779067] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Acquired lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 656.779067] env[65121]: DEBUG nova.network.neutron [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Refreshing network info cache for port b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 656.859242] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106058, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.873133] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "refresh_cache-7a0d0769-31c5-4ae1-8520-a744f64d39c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.873483] env[65121]: DEBUG nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Instance network_info: |[{"id": "33c6c3ae-feb6-4512-ac82-46c27afad318", "address": "fa:16:3e:20:ab:a5", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33c6c3ae-fe", "ovs_interfaceid": "33c6c3ae-feb6-4512-ac82-46c27afad318", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 656.874191] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:ab:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33c6c3ae-feb6-4512-ac82-46c27afad318', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 656.885141] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Creating folder: Project (b35af4e2b6844a9081f34466661c83b2). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 656.886859] env[65121]: DEBUG nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 656.889663] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-625c75e4-94fe-41f0-85ad-070ec42fcdb3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.912359] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Created folder: Project (b35af4e2b6844a9081f34466661c83b2) in parent group-v993268. [ 656.912359] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Creating folder: Instances. Parent ref: group-v993325. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 656.912917] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83da8a6c-52e9-432a-af17-89f808709189 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.927808] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Created folder: Instances in parent group-v993325. [ 656.928708] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 656.931456] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 656.932215] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 656.932215] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 656.932215] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 656.932598] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 656.933227] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 656.933487] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.933884] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 656.934106] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 656.935041] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 656.935330] env[65121]: DEBUG nova.virt.hardware [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 656.935945] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 656.936595] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831b8aaf-5064-422a-9844-51504d3aa0af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.941023] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34a1ab00-5a53-442c-8f61-fd7cb988ccab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.970057] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76fc824-e7fb-4b5f-8bae-57644980e6e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.975644] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 656.975644] env[65121]: value = "task-5106062" [ 656.975644] env[65121]: _type = "Task" [ 656.975644] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.998088] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106062, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.027417] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106057, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.766591} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.027832] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 6a4841b6-6076-4e9c-a6cf-a658957e2931/6a4841b6-6076-4e9c-a6cf-a658957e2931.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 657.028061] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 657.028319] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-915b74a9-1068-431d-b33b-242f9d85c17d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.038347] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 657.038347] env[65121]: value = "task-5106063" [ 657.038347] env[65121]: _type = "Task" [ 657.038347] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.048613] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106063, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.067328] env[65121]: DEBUG nova.scheduler.client.report [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 657.073495] env[65121]: INFO nova.compute.manager [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Took 32.09 seconds to build instance. [ 657.204221] env[65121]: DEBUG oslo_vmware.api [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5106059, 'name': PowerOffVM_Task, 'duration_secs': 0.248797} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.204514] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 657.204910] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 657.205234] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8691974e-a57e-4858-a218-929fdcb497e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.238620] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 657.238829] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 657.239103] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleting the datastore file [datastore2] 20396b81-2041-4bf2-85c4-8ea66c838ec4 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 657.239354] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfdef66e-5a89-4eea-bbd7-c3c5b17b26b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.247879] env[65121]: DEBUG oslo_vmware.api [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for the task: (returnval){ [ 657.247879] env[65121]: value = "task-5106065" [ 657.247879] env[65121]: _type = "Task" [ 657.247879] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.261259] env[65121]: DEBUG oslo_vmware.api [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5106065, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.285269] env[65121]: WARNING neutronclient.v2_0.client [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.286263] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.286733] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.357730] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106058, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.490771] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106062, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.532541] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Successfully created port: 3666dad5-ffa7-4c68-88b9-c2cb9d318615 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 657.548775] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106063, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.280792} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.549159] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 657.550042] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4efe0b79-d700-4d6a-b56e-157e7e04e8d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.555914] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.556288] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.579617] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.333s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.580291] env[65121]: DEBUG nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 657.597580] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 6a4841b6-6076-4e9c-a6cf-a658957e2931/6a4841b6-6076-4e9c-a6cf-a658957e2931.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 657.599073] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.964s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.599073] env[65121]: DEBUG nova.objects.instance [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lazy-loading 'resources' on Instance uuid bad21c0b-6534-46b5-97a3-a26e5ecb5c96 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 657.600350] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4514ca0-8763-46a5-98f0-2e279aaa0b52 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.598s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.602029] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21d1abcd-602b-4502-99e1-434185cc9dc1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.628483] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 657.628483] env[65121]: value = "task-5106066" [ 657.628483] env[65121]: _type = "Task" [ 657.628483] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.642155] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106066, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.726522] env[65121]: DEBUG nova.compute.manager [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Received event network-changed-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 657.726751] env[65121]: DEBUG nova.compute.manager [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Refreshing instance network info cache due to event network-changed-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 657.727016] env[65121]: DEBUG oslo_concurrency.lockutils [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Acquiring lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.727069] env[65121]: DEBUG oslo_concurrency.lockutils [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Acquired lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.727218] env[65121]: DEBUG nova.network.neutron [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Refreshing network info cache for port ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 657.769190] env[65121]: DEBUG oslo_vmware.api [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Task: {'id': task-5106065, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169322} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.769562] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 657.769844] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 657.770079] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.770292] env[65121]: INFO nova.compute.manager [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Took 1.11 seconds to destroy the instance on the hypervisor. [ 657.770590] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 657.771634] env[65121]: DEBUG nova.compute.manager [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 657.771634] env[65121]: DEBUG nova.network.neutron [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 657.771634] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.771866] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.772507] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.862506] env[65121]: DEBUG oslo_vmware.api [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106058, 'name': RemoveSnapshot_Task, 'duration_secs': 1.350715} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.862934] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 657.863101] env[65121]: INFO nova.compute.manager [None req-c3b8491e-2f03-4ade-8b21-54d06ee2c0d5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Took 17.42 seconds to snapshot the instance on the hypervisor. [ 657.998964] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106062, 'name': CreateVM_Task, 'duration_secs': 0.608895} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.999261] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 657.999761] env[65121]: WARNING neutronclient.v2_0.client [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.000218] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.000462] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 658.000803] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 658.001140] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef157bae-e507-4b23-8e49-ba11189b6362 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.010728] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 658.010728] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ee84d7-347f-441f-5247-4b66e1d4d53a" [ 658.010728] env[65121]: _type = "Task" [ 658.010728] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.020710] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ee84d7-347f-441f-5247-4b66e1d4d53a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.103853] env[65121]: DEBUG nova.compute.utils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 658.103853] env[65121]: DEBUG nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 658.104335] env[65121]: DEBUG nova.network.neutron [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 658.105169] env[65121]: WARNING neutronclient.v2_0.client [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.105596] env[65121]: WARNING neutronclient.v2_0.client [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.106389] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.106853] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.141799] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7b2094-5056-4335-b95d-3f0c64aca72e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.148507] env[65121]: DEBUG nova.network.neutron [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 658.149095] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.157808] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106066, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.159092] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca94ebeb-10bf-4276-954b-1afefcbe1696 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.198741] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea53db0-6f0d-48cb-b6e0-79bc670548f1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.211036] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab0ecc6-478b-415f-a496-cf216c0ecfa6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.226520] env[65121]: DEBUG nova.compute.provider_tree [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.233041] env[65121]: WARNING neutronclient.v2_0.client [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.233728] env[65121]: WARNING openstack [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.234129] env[65121]: WARNING openstack [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.262614] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Successfully created port: 142f7620-365e-4b15-8278-d8cee47cfccb {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 658.313631] env[65121]: WARNING neutronclient.v2_0.client [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.314444] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.314800] env[65121]: WARNING openstack [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.461738] env[65121]: DEBUG nova.policy [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfa7595ebb584044a2d0e8f3e469e9b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e92570886f544d9ea9fe0947e8e93d9c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 658.477128] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.477509] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.524894] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ee84d7-347f-441f-5247-4b66e1d4d53a, 'name': SearchDatastore_Task, 'duration_secs': 0.015596} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.525233] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.525457] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 658.525686] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.525823] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 658.525997] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 658.526321] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-827698c1-80e9-4ba5-b8d2-b57a5daad850 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.537686] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 658.537686] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 658.538251] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fd3009b-b1b7-4009-b376-ed1aac41b3f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.545681] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 658.545681] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5298c4bc-687e-114b-cd83-8033f9779268" [ 658.545681] env[65121]: _type = "Task" [ 658.545681] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.559535] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5298c4bc-687e-114b-cd83-8033f9779268, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.619125] env[65121]: DEBUG nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 658.642676] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106066, 'name': ReconfigVM_Task, 'duration_secs': 0.557232} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.643569] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 6a4841b6-6076-4e9c-a6cf-a658957e2931/6a4841b6-6076-4e9c-a6cf-a658957e2931.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 658.644431] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b54d513-cee7-45dc-9260-e61b4693a31f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.651923] env[65121]: DEBUG nova.network.neutron [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 658.653817] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 658.653817] env[65121]: value = "task-5106067" [ 658.653817] env[65121]: _type = "Task" [ 658.653817] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.666981] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106067, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.730647] env[65121]: DEBUG nova.scheduler.client.report [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 658.869686] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Successfully created port: 1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 659.060161] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5298c4bc-687e-114b-cd83-8033f9779268, 'name': SearchDatastore_Task, 'duration_secs': 0.014321} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.061056] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60772343-9d59-4b5f-a69b-3576e07ee1fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.068530] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 659.068530] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52eefe3a-00ab-c6d9-6851-65c5d114a3fa" [ 659.068530] env[65121]: _type = "Task" [ 659.068530] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.078819] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52eefe3a-00ab-c6d9-6851-65c5d114a3fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.103041] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 659.103041] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.103041] env[65121]: DEBUG nova.compute.manager [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Going to confirm migration 1 {{(pid=65121) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 659.162756] env[65121]: INFO nova.compute.manager [-] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Took 1.39 seconds to deallocate network for instance. [ 659.177669] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106067, 'name': Rename_Task, 'duration_secs': 0.210487} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.179723] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 659.179795] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5daba576-ae25-486d-aeca-521b20ebc6ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.190830] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 659.190830] env[65121]: value = "task-5106068" [ 659.190830] env[65121]: _type = "Task" [ 659.190830] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.200984] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106068, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.240373] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.642s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.254055] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.610s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.260479] env[65121]: INFO nova.compute.claims [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.265223] env[65121]: DEBUG nova.network.neutron [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Successfully created port: 1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 659.299405] env[65121]: INFO nova.scheduler.client.report [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted allocations for instance bad21c0b-6534-46b5-97a3-a26e5ecb5c96 [ 659.353246] env[65121]: DEBUG nova.network.neutron [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Updating instance_info_cache with network_info: [{"id": "a27cb922-5612-41b7-9637-f2c0bff9c8b5", "address": "fa:16:3e:ba:1a:a2", "network": {"id": "e62fba86-0632-4afa-9490-1123f16bd48e", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-744677946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b816d77d6f3b4bc7a20d5b97ba068006", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa27cb922-56", "ovs_interfaceid": "a27cb922-5612-41b7-9637-f2c0bff9c8b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 659.411745] env[65121]: WARNING neutronclient.v2_0.client [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.412505] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.413154] env[65121]: WARNING openstack [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.584335] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52eefe3a-00ab-c6d9-6851-65c5d114a3fa, 'name': SearchDatastore_Task, 'duration_secs': 0.026448} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.584780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.585155] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 7a0d0769-31c5-4ae1-8520-a744f64d39c6/7a0d0769-31c5-4ae1-8520-a744f64d39c6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 659.585449] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7180cc28-8e7c-420d-8d04-eb1a5e71c0d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.593577] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 659.593577] env[65121]: value = "task-5106069" [ 659.593577] env[65121]: _type = "Task" [ 659.593577] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.605666] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106069, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.609922] env[65121]: WARNING neutronclient.v2_0.client [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.630674] env[65121]: DEBUG nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 659.670120] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 659.670120] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 659.670523] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 659.670953] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 659.671249] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 659.672022] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 659.672022] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.672022] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 659.672022] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 659.672376] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 659.672601] env[65121]: DEBUG nova.virt.hardware [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 659.673724] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d42f764-a64e-43ea-a1a8-dc5a51fd5848 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.680221] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 659.684430] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c1f443-446c-4a88-b3e5-1942e41de6f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.712806] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106068, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.817793] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f1c1971a-25c7-49be-be6e-3018fc07749a tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "bad21c0b-6534-46b5-97a3-a26e5ecb5c96" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.154s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.857845] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Releasing lock "refresh_cache-6218957b-6329-4004-97ca-07231b55937a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.858585] env[65121]: DEBUG nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Instance network_info: |[{"id": "a27cb922-5612-41b7-9637-f2c0bff9c8b5", "address": "fa:16:3e:ba:1a:a2", "network": {"id": "e62fba86-0632-4afa-9490-1123f16bd48e", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-744677946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b816d77d6f3b4bc7a20d5b97ba068006", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa27cb922-56", "ovs_interfaceid": "a27cb922-5612-41b7-9637-f2c0bff9c8b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 659.859873] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:1a:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4223acd2-30f7-440e-b975-60b30d931694', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a27cb922-5612-41b7-9637-f2c0bff9c8b5', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 659.876320] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Creating folder: Project (b816d77d6f3b4bc7a20d5b97ba068006). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 659.876966] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7af6a1f9-bfe8-4d8f-8f14-0d9ec10cd7c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.901421] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Created folder: Project (b816d77d6f3b4bc7a20d5b97ba068006) in parent group-v993268. [ 659.901421] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Creating folder: Instances. Parent ref: group-v993328. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 659.901421] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1883df45-9cb8-4013-a752-223df39f1d5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.917633] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Created folder: Instances in parent group-v993328. [ 659.917633] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 659.917633] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6218957b-6329-4004-97ca-07231b55937a] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 659.917633] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b9fa8b9-6a9c-4256-a4e8-d68f906420f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.946759] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 659.946759] env[65121]: value = "task-5106072" [ 659.946759] env[65121]: _type = "Task" [ 659.946759] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.959164] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106072, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.106476] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106069, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.203655] env[65121]: DEBUG oslo_vmware.api [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106068, 'name': PowerOnVM_Task, 'duration_secs': 0.797725} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.204073] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 660.204341] env[65121]: INFO nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Took 12.96 seconds to spawn the instance on the hypervisor. [ 660.204537] env[65121]: DEBUG nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 660.205506] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d8bf76-4ab3-40a9-bdfd-a1c598276358 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.460544] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106072, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.607200] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106069, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.661217} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.610121] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 7a0d0769-31c5-4ae1-8520-a744f64d39c6/7a0d0769-31c5-4ae1-8520-a744f64d39c6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 660.610348] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 660.610864] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0bb66a1-140e-4662-8829-fdbfa47ae685 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.619581] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 660.619581] env[65121]: value = "task-5106073" [ 660.619581] env[65121]: _type = "Task" [ 660.619581] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.634046] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106073, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.651834] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5950c852-7786-4377-a83f-f365bab465b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.661327] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2389f76-2cd0-43eb-b573-5c84d9f22a78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.698629] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce680e9a-8cb5-45c6-a1be-cef9ee89c4f1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.707211] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1098c6b-d586-43c1-91fb-f691e40248e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.731440] env[65121]: DEBUG nova.compute.provider_tree [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.735112] env[65121]: INFO nova.compute.manager [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Took 34.73 seconds to build instance. [ 660.954045] env[65121]: DEBUG nova.network.neutron [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Successfully updated port: 1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 660.962515] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106072, 'name': CreateVM_Task, 'duration_secs': 0.625013} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.963067] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6218957b-6329-4004-97ca-07231b55937a] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 660.963273] env[65121]: WARNING neutronclient.v2_0.client [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.963635] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.963781] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.964673] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 660.964673] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78aa9532-6874-4f39-aed0-ee858b24b6d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.972239] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 660.972239] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5244a392-414b-0e60-7f12-341c538d6050" [ 660.972239] env[65121]: _type = "Task" [ 660.972239] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.976251] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Successfully updated port: 3666dad5-ffa7-4c68-88b9-c2cb9d318615 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 660.983475] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5244a392-414b-0e60-7f12-341c538d6050, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.001097] env[65121]: WARNING openstack [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.001481] env[65121]: WARNING openstack [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.132533] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Received event network-vif-plugged-18509dce-9df7-41cc-af3e-70f419fbb5d5 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 661.132634] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Acquiring lock "6a4841b6-6076-4e9c-a6cf-a658957e2931-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.132896] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.132979] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.133160] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] No waiting events found dispatching network-vif-plugged-18509dce-9df7-41cc-af3e-70f419fbb5d5 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 661.133429] env[65121]: WARNING nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Received unexpected event network-vif-plugged-18509dce-9df7-41cc-af3e-70f419fbb5d5 for instance with vm_state active and task_state None. [ 661.133477] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Received event network-changed-18509dce-9df7-41cc-af3e-70f419fbb5d5 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 661.133590] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Refreshing instance network info cache due to event network-changed-18509dce-9df7-41cc-af3e-70f419fbb5d5. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 661.133847] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Acquiring lock "refresh_cache-6a4841b6-6076-4e9c-a6cf-a658957e2931" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.134015] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Acquired lock "refresh_cache-6a4841b6-6076-4e9c-a6cf-a658957e2931" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.134163] env[65121]: DEBUG nova.network.neutron [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Refreshing network info cache for port 18509dce-9df7-41cc-af3e-70f419fbb5d5 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 661.146359] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106073, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076118} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.148655] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 661.149831] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a06ec2-1869-422c-8ead-3c0d4ce1e8b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.179924] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Reconfiguring VM instance instance-00000013 to attach disk [datastore2] 7a0d0769-31c5-4ae1-8520-a744f64d39c6/7a0d0769-31c5-4ae1-8520-a744f64d39c6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 661.185958] env[65121]: WARNING neutronclient.v2_0.client [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.186093] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.186233] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.186412] env[65121]: DEBUG nova.network.neutron [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 661.186600] env[65121]: DEBUG nova.objects.instance [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lazy-loading 'info_cache' on Instance uuid 5f6e4f46-0745-42c0-a779-6cffb60e21ca {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 661.191025] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-216c3844-3f93-41d0-b9e0-608a5ada67b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.212855] env[65121]: DEBUG nova.network.neutron [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Updated VIF entry in instance network info cache for port b25aa25b-2d1b-414e-883a-324894d4b483. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 661.213148] env[65121]: DEBUG nova.network.neutron [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Updating instance_info_cache with network_info: [{"id": "b25aa25b-2d1b-414e-883a-324894d4b483", "address": "fa:16:3e:0e:be:f6", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25aa25b-2d", "ovs_interfaceid": "b25aa25b-2d1b-414e-883a-324894d4b483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 661.216586] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "6887bb66-7186-4d58-be59-d0c53c3221c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.216813] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "6887bb66-7186-4d58-be59-d0c53c3221c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.217016] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "6887bb66-7186-4d58-be59-d0c53c3221c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.217188] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "6887bb66-7186-4d58-be59-d0c53c3221c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.220021] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "6887bb66-7186-4d58-be59-d0c53c3221c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.221160] env[65121]: INFO nova.compute.manager [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Terminating instance [ 661.230410] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 661.230410] env[65121]: value = "task-5106074" [ 661.230410] env[65121]: _type = "Task" [ 661.230410] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.236816] env[65121]: DEBUG nova.scheduler.client.report [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 661.242997] env[65121]: DEBUG oslo_concurrency.lockutils [None req-080dc3b7-450b-48a9-8bee-693d129eeb55 tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.212s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.248988] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106074, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.269510] env[65121]: WARNING neutronclient.v2_0.client [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.270277] env[65121]: WARNING openstack [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.270768] env[65121]: WARNING openstack [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.439809] env[65121]: DEBUG nova.network.neutron [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Updated VIF entry in instance network info cache for port ec7592a3-0a2f-43b5-9c55-2e59d85c17cf. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 661.439809] env[65121]: DEBUG nova.network.neutron [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Updating instance_info_cache with network_info: [{"id": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "address": "fa:16:3e:6b:03:cf", "network": {"id": "81cf3ab5-bc0a-4335-88c1-87fc12584925", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-98268708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "def20b65ac8341c1a60a747260589d5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec7592a3-0a", "ovs_interfaceid": "ec7592a3-0a2f-43b5-9c55-2e59d85c17cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 661.465524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.465524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.465524] env[65121]: DEBUG nova.network.neutron [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 661.486131] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5244a392-414b-0e60-7f12-341c538d6050, 'name': SearchDatastore_Task, 'duration_secs': 0.01434} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.488100] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.488100] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 661.488100] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.488100] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.488372] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 661.488372] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58838d2d-87fa-4976-8e25-5a1b9218267e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.500187] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 661.500409] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 661.501601] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b31edc39-9e3d-4e24-b22f-8af6bbe6018a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.509981] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 661.509981] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5248c02d-379b-5b2c-6364-6c1ecfb071ee" [ 661.509981] env[65121]: _type = "Task" [ 661.509981] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.522493] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5248c02d-379b-5b2c-6364-6c1ecfb071ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.646024] env[65121]: WARNING neutronclient.v2_0.client [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.646484] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.646828] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.725468] env[65121]: DEBUG oslo_concurrency.lockutils [req-31ae82d9-de3d-4b0a-976a-1fc3da3b0405 req-1dba8395-1481-46a2-872b-18d4a69bb71c service nova] Releasing lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.729070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "refresh_cache-6887bb66-7186-4d58-be59-d0c53c3221c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.729261] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquired lock "refresh_cache-6887bb66-7186-4d58-be59-d0c53c3221c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.729431] env[65121]: DEBUG nova.network.neutron [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 661.745913] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.746465] env[65121]: DEBUG nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 661.749366] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106074, 'name': ReconfigVM_Task, 'duration_secs': 0.318527} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.750590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.500s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.752086] env[65121]: INFO nova.compute.claims [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.755076] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Reconfigured VM instance instance-00000013 to attach disk [datastore2] 7a0d0769-31c5-4ae1-8520-a744f64d39c6/7a0d0769-31c5-4ae1-8520-a744f64d39c6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 661.756902] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42fd9b3f-2963-4040-89c0-36209a194cb1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.767541] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 661.767541] env[65121]: value = "task-5106075" [ 661.767541] env[65121]: _type = "Task" [ 661.767541] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.781063] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106075, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.902152] env[65121]: DEBUG oslo_concurrency.lockutils [req-50f0f48c-f18c-4667-ba0c-e267d531c914 req-8ac73e84-a3fc-4a6a-8b7b-41ad2017f954 service nova] Releasing lock "refresh_cache-4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.968912] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.969444] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.021231] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5248c02d-379b-5b2c-6364-6c1ecfb071ee, 'name': SearchDatastore_Task, 'duration_secs': 0.016949} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.022125] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b466458c-5f4a-4e5f-b8eb-3179d5d5562d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.027924] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 662.027924] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dd2019-5392-4913-58cb-4f50729e8a3e" [ 662.027924] env[65121]: _type = "Task" [ 662.027924] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.039405] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dd2019-5392-4913-58cb-4f50729e8a3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.227056] env[65121]: WARNING neutronclient.v2_0.client [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.227767] env[65121]: WARNING openstack [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.228229] env[65121]: WARNING openstack [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.236782] env[65121]: WARNING neutronclient.v2_0.client [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.238057] env[65121]: WARNING openstack [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.238057] env[65121]: WARNING openstack [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.252338] env[65121]: DEBUG nova.compute.utils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 662.253780] env[65121]: DEBUG nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 662.254007] env[65121]: DEBUG nova.network.neutron [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 662.255523] env[65121]: WARNING neutronclient.v2_0.client [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.255994] env[65121]: WARNING neutronclient.v2_0.client [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.259062] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.259062] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.279361] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106075, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.336167] env[65121]: DEBUG nova.network.neutron [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 662.385221] env[65121]: DEBUG nova.network.neutron [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 662.426335] env[65121]: DEBUG nova.policy [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8c4e9798224438c874051d6f2e42966', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec5741daa3f3411bb74b3b965bbeef44', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 662.505658] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.506125] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.540263] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dd2019-5392-4913-58cb-4f50729e8a3e, 'name': SearchDatastore_Task, 'duration_secs': 0.020456} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.540659] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 662.540834] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 6218957b-6329-4004-97ca-07231b55937a/6218957b-6329-4004-97ca-07231b55937a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 662.541229] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-572331f6-553f-4a23-92bb-da089a677732 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.550669] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 662.550669] env[65121]: value = "task-5106076" [ 662.550669] env[65121]: _type = "Task" [ 662.550669] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.567972] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.765813] env[65121]: DEBUG nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 662.788776] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.789197] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.804066] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106075, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.885708] env[65121]: WARNING openstack [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.887348] env[65121]: WARNING openstack [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.941798] env[65121]: DEBUG nova.network.neutron [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.065203] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106076, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.241050] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Successfully updated port: 142f7620-365e-4b15-8278-d8cee47cfccb {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 663.297958] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106075, 'name': Rename_Task, 'duration_secs': 1.161767} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.301330] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 663.302858] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea9de7fc-b9ae-41a0-8adf-94899e2667ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.305940] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b4de9c-dc55-455e-b1d6-61c6884c67a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.317276] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88280ca9-4a97-42bf-a44f-00da1c245ed8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.320981] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 663.320981] env[65121]: value = "task-5106077" [ 663.320981] env[65121]: _type = "Task" [ 663.320981] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.353991] env[65121]: DEBUG nova.network.neutron [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Successfully created port: 813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 663.357321] env[65121]: WARNING neutronclient.v2_0.client [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.357967] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.359217] env[65121]: WARNING openstack [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.367205] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfb4824-0594-4b74-89c0-921ebc7830a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.375717] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106077, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.386484] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dea48d4-40f9-47ff-af7b-b72f3cf3b92e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.406499] env[65121]: DEBUG nova.compute.provider_tree [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.425985] env[65121]: WARNING neutronclient.v2_0.client [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.427100] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.427700] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.446388] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Releasing lock "refresh_cache-6887bb66-7186-4d58-be59-d0c53c3221c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.446833] env[65121]: DEBUG nova.compute.manager [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 663.447088] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.447982] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bd8e5b-8667-45a0-b70f-f78cd179cd7e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.457422] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 663.457766] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15e2fe45-8a20-4233-9ebb-1750f91e195f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.466883] env[65121]: DEBUG oslo_vmware.api [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 663.466883] env[65121]: value = "task-5106078" [ 663.466883] env[65121]: _type = "Task" [ 663.466883] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.477273] env[65121]: DEBUG oslo_vmware.api [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5106078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.520772] env[65121]: WARNING neutronclient.v2_0.client [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.521617] env[65121]: WARNING openstack [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.522360] env[65121]: WARNING openstack [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.567028] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592442} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.567372] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 6218957b-6329-4004-97ca-07231b55937a/6218957b-6329-4004-97ca-07231b55937a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 663.567615] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 663.568015] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-154338a0-a6e4-457b-92ad-af1631dc76a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.577209] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 663.577209] env[65121]: value = "task-5106079" [ 663.577209] env[65121]: _type = "Task" [ 663.577209] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.587962] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106079, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.642744] env[65121]: DEBUG nova.network.neutron [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updating instance_info_cache with network_info: [{"id": "1c577738-fcc1-4754-9cee-519ac385ae80", "address": "fa:16:3e:ed:26:a9", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c577738-fc", "ovs_interfaceid": "1c577738-fcc1-4754-9cee-519ac385ae80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.791842] env[65121]: DEBUG nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 663.819710] env[65121]: DEBUG nova.network.neutron [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Updated VIF entry in instance network info cache for port 18509dce-9df7-41cc-af3e-70f419fbb5d5. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 663.820210] env[65121]: DEBUG nova.network.neutron [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Updating instance_info_cache with network_info: [{"id": "18509dce-9df7-41cc-af3e-70f419fbb5d5", "address": "fa:16:3e:3e:13:c2", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.95", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18509dce-9d", "ovs_interfaceid": "18509dce-9df7-41cc-af3e-70f419fbb5d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.833537] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106077, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.836466] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 663.836772] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 663.836954] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 663.837159] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 663.837299] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 663.837434] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 663.837725] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.837875] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 663.838042] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 663.838195] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 663.838355] env[65121]: DEBUG nova.virt.hardware [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 663.839564] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f780a23a-6faa-4644-85f7-c8fd67f93e9b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.848996] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d289c1d1-79c3-45fb-988a-765c65c622fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.910351] env[65121]: DEBUG nova.scheduler.client.report [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 663.918337] env[65121]: DEBUG nova.network.neutron [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance_info_cache with network_info: [{"id": "fc77c1b5-9d9b-479a-a553-e7961313af08", "address": "fa:16:3e:f6:e5:10", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.52", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc77c1b5-9d", "ovs_interfaceid": "fc77c1b5-9d9b-479a-a553-e7961313af08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.981878] env[65121]: DEBUG oslo_vmware.api [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5106078, 'name': PowerOffVM_Task, 'duration_secs': 0.221471} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.982312] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 663.982577] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 663.982963] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8cfe913-2ac3-4fb2-8c9d-e04c5bd766e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.015679] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 664.016406] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 664.016799] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Deleting the datastore file [datastore1] 6887bb66-7186-4d58-be59-d0c53c3221c6 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 664.017333] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-70f23b19-7891-470e-8e9d-70beb3d6e2cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.026375] env[65121]: DEBUG oslo_vmware.api [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for the task: (returnval){ [ 664.026375] env[65121]: value = "task-5106081" [ 664.026375] env[65121]: _type = "Task" [ 664.026375] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.037738] env[65121]: DEBUG oslo_vmware.api [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5106081, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.088583] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106079, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109558} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.089066] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 664.089671] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc15420-2547-407a-84d7-c443dc8facb3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.116427] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] 6218957b-6329-4004-97ca-07231b55937a/6218957b-6329-4004-97ca-07231b55937a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 664.116764] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4646f320-0524-4176-ad62-cc5cacc470b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.141271] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 664.141271] env[65121]: value = "task-5106082" [ 664.141271] env[65121]: _type = "Task" [ 664.141271] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.149326] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Releasing lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.149667] env[65121]: DEBUG nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance network_info: |[{"id": "1c577738-fcc1-4754-9cee-519ac385ae80", "address": "fa:16:3e:ed:26:a9", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c577738-fc", "ovs_interfaceid": "1c577738-fcc1-4754-9cee-519ac385ae80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 664.149996] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "c2f9fa49-4273-4885-b59d-0635ed739d65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.150234] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.151447] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:26:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c577738-fcc1-4754-9cee-519ac385ae80', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 664.160182] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 664.161759] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 664.162098] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dccd63c0-15f8-4eea-a66a-27b0490bbbc8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.185753] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106082, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.192903] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 664.192903] env[65121]: value = "task-5106083" [ 664.192903] env[65121]: _type = "Task" [ 664.192903] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.203103] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106083, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.328398] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Releasing lock "refresh_cache-6a4841b6-6076-4e9c-a6cf-a658957e2931" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.328676] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Received event network-vif-plugged-33c6c3ae-feb6-4512-ac82-46c27afad318 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 664.328886] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Acquiring lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.329112] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.329288] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.329472] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] No waiting events found dispatching network-vif-plugged-33c6c3ae-feb6-4512-ac82-46c27afad318 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 664.329660] env[65121]: WARNING nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Received unexpected event network-vif-plugged-33c6c3ae-feb6-4512-ac82-46c27afad318 for instance with vm_state building and task_state spawning. [ 664.329818] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Received event network-changed-33c6c3ae-feb6-4512-ac82-46c27afad318 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 664.329963] env[65121]: DEBUG nova.compute.manager [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Refreshing instance network info cache due to event network-changed-33c6c3ae-feb6-4512-ac82-46c27afad318. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 664.330169] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Acquiring lock "refresh_cache-7a0d0769-31c5-4ae1-8520-a744f64d39c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.330298] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Acquired lock "refresh_cache-7a0d0769-31c5-4ae1-8520-a744f64d39c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 664.330497] env[65121]: DEBUG nova.network.neutron [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Refreshing network info cache for port 33c6c3ae-feb6-4512-ac82-46c27afad318 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 664.339487] env[65121]: DEBUG oslo_vmware.api [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106077, 'name': PowerOnVM_Task, 'duration_secs': 0.756865} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.339952] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 664.340293] env[65121]: INFO nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Took 13.48 seconds to spawn the instance on the hypervisor. [ 664.341503] env[65121]: DEBUG nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 664.343495] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae51f80-3b95-4c5e-9d70-4979136369b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.416708] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.666s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.417624] env[65121]: DEBUG nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 664.422847] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.256s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.423478] env[65121]: DEBUG nova.objects.instance [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lazy-loading 'resources' on Instance uuid 5e915c8e-150a-4bfd-8005-a7a1a119057d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 664.425878] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-5f6e4f46-0745-42c0-a779-6cffb60e21ca" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.426116] env[65121]: DEBUG nova.objects.instance [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lazy-loading 'migration_context' on Instance uuid 5f6e4f46-0745-42c0-a779-6cffb60e21ca {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 664.434089] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cde36-1906-eedc-7bdb-0641999b5da4/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 664.436483] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae5bc73-ea44-4213-ada1-cb9deb4bb558 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.445051] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cde36-1906-eedc-7bdb-0641999b5da4/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 664.445215] env[65121]: ERROR oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cde36-1906-eedc-7bdb-0641999b5da4/disk-0.vmdk due to incomplete transfer. [ 664.445460] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4fca23b1-9b3a-42d5-98f7-aea978f9f42d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.455014] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528cde36-1906-eedc-7bdb-0641999b5da4/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 664.455347] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Uploaded image 8af4d2e1-314a-42f6-b6eb-dfa13df15f20 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 664.457330] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 664.457630] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-adb83b56-f1f4-4471-96c1-42f64db77c09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.468316] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 664.468316] env[65121]: value = "task-5106084" [ 664.468316] env[65121]: _type = "Task" [ 664.468316] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.478037] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106084, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.541843] env[65121]: DEBUG oslo_vmware.api [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Task: {'id': task-5106081, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228527} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.542189] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 664.542450] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 664.542699] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.542936] env[65121]: INFO nova.compute.manager [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Took 1.10 seconds to destroy the instance on the hypervisor. [ 664.543323] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 664.543597] env[65121]: DEBUG nova.compute.manager [-] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 664.543734] env[65121]: DEBUG nova.network.neutron [-] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 664.544112] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.544721] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.545038] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.653198] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106082, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.661812] env[65121]: DEBUG nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 664.703170] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106083, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.834286] env[65121]: WARNING neutronclient.v2_0.client [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.834989] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.835419] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.869205] env[65121]: INFO nova.compute.manager [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Took 31.65 seconds to build instance. [ 664.922116] env[65121]: DEBUG nova.compute.utils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 664.924608] env[65121]: DEBUG nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 664.924894] env[65121]: DEBUG nova.network.neutron [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 664.925337] env[65121]: WARNING neutronclient.v2_0.client [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.925566] env[65121]: WARNING neutronclient.v2_0.client [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.926181] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.926498] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.940189] env[65121]: DEBUG nova.objects.base [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Object Instance<5f6e4f46-0745-42c0-a779-6cffb60e21ca> lazy-loaded attributes: info_cache,migration_context {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 664.943454] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c76bc8-9a51-413e-888c-e7a926798f9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.974143] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27b52282-d92d-472e-8a76-71d02a1c4384 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.982124] env[65121]: DEBUG oslo_vmware.api [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 664.982124] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528ded8d-ca72-2322-083b-bcc2304f34d1" [ 664.982124] env[65121]: _type = "Task" [ 664.982124] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.986522] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106084, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.998422] env[65121]: DEBUG oslo_vmware.api [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528ded8d-ca72-2322-083b-bcc2304f34d1, 'name': SearchDatastore_Task, 'duration_secs': 0.011883} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.998936] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.088203] env[65121]: DEBUG nova.network.neutron [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Successfully updated port: 813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 665.156059] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106082, 'name': ReconfigVM_Task, 'duration_secs': 0.730564} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.156449] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Reconfigured VM instance instance-00000014 to attach disk [datastore1] 6218957b-6329-4004-97ca-07231b55937a/6218957b-6329-4004-97ca-07231b55937a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 665.157359] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f68e54e1-b371-4d7a-9250-3a87844af8df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.164757] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 665.164757] env[65121]: value = "task-5106085" [ 665.164757] env[65121]: _type = "Task" [ 665.164757] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.177669] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106085, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.183607] env[65121]: DEBUG nova.network.neutron [-] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 665.183791] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.187348] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.204536] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106083, 'name': CreateVM_Task, 'duration_secs': 0.625392} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.204630] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 665.205353] env[65121]: WARNING neutronclient.v2_0.client [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.205525] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.206532] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.206532] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 665.206532] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed425bda-73ef-40e0-80c1-3260268a70af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.216022] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 665.216022] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e7e5bb-f41a-985e-9656-a9be75ace0ee" [ 665.216022] env[65121]: _type = "Task" [ 665.216022] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.226896] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e7e5bb-f41a-985e-9656-a9be75ace0ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.372277] env[65121]: DEBUG oslo_concurrency.lockutils [None req-66ae4236-f803-4b6b-9623-8567ccef315e tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.167s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.413969] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39761260-2dc8-4af1-b3fa-f4ab6edd2859 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.421849] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd632ac-2506-4be5-8205-1d9c7796c51a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.430453] env[65121]: DEBUG nova.policy [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c21a3c41d3d64ef4aff6ab64ad2c8d7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5724c072955c4cd3a5500dedf3b972a3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 665.438435] env[65121]: DEBUG nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 665.475327] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.475327] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.486521] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5ce714-3442-4f66-a218-722b80cb708c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.503465] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106084, 'name': Destroy_Task, 'duration_secs': 0.678881} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.506281] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Destroyed the VM [ 665.506537] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 665.506857] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-edea8592-5dba-4ef5-b6a9-a4a6a57d03c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.510021] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aca00a5-53a9-42b4-bfd6-cb82656d68f2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.529607] env[65121]: DEBUG nova.compute.provider_tree [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.533377] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 665.533377] env[65121]: value = "task-5106086" [ 665.533377] env[65121]: _type = "Task" [ 665.533377] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.552471] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106086, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.593957] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.594227] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquired lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.594354] env[65121]: DEBUG nova.network.neutron [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 665.644220] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Successfully updated port: 1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 665.676495] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106085, 'name': Rename_Task, 'duration_secs': 0.368702} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.676753] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 665.677100] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71d8a3ae-07b5-4356-86c6-f073df00a018 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.686546] env[65121]: DEBUG nova.network.neutron [-] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.691430] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 665.691430] env[65121]: value = "task-5106087" [ 665.691430] env[65121]: _type = "Task" [ 665.691430] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.702646] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106087, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.727388] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e7e5bb-f41a-985e-9656-a9be75ace0ee, 'name': SearchDatastore_Task, 'duration_secs': 0.010717} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.727723] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 665.727958] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 665.728599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.728599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.728599] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 665.729560] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d928672-997e-4422-ad86-d56f04954d76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.744651] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 665.744863] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 665.746151] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-820d0f3e-20fc-4de0-86c4-244e06f3685b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.752841] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 665.752841] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c6fccc-ae3d-67f7-6de2-35febbbbbe61" [ 665.752841] env[65121]: _type = "Task" [ 665.752841] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.763638] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c6fccc-ae3d-67f7-6de2-35febbbbbe61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.936465] env[65121]: WARNING neutronclient.v2_0.client [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.937400] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.938688] env[65121]: WARNING openstack [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.953994] env[65121]: DEBUG nova.network.neutron [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Successfully created port: f3966bcc-2651-436f-8bcb-da2f0b915c10 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 666.040263] env[65121]: DEBUG nova.scheduler.client.report [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 666.056793] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106086, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.070598] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "27940143-16b5-4263-b23c-354ed8ea8866" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.071327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "27940143-16b5-4263-b23c-354ed8ea8866" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.098172] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.098602] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.148667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.148858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.149019] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 666.190753] env[65121]: INFO nova.compute.manager [-] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Took 1.65 seconds to deallocate network for instance. [ 666.207094] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106087, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.211761] env[65121]: DEBUG nova.network.neutron [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 666.254764] env[65121]: DEBUG nova.network.neutron [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Updated VIF entry in instance network info cache for port 33c6c3ae-feb6-4512-ac82-46c27afad318. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 666.255216] env[65121]: DEBUG nova.network.neutron [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Updating instance_info_cache with network_info: [{"id": "33c6c3ae-feb6-4512-ac82-46c27afad318", "address": "fa:16:3e:20:ab:a5", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33c6c3ae-fe", "ovs_interfaceid": "33c6c3ae-feb6-4512-ac82-46c27afad318", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 666.268166] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c6fccc-ae3d-67f7-6de2-35febbbbbe61, 'name': SearchDatastore_Task, 'duration_secs': 0.017435} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.269679] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-885b85cc-e1aa-40a0-ae17-94d63e01a8f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.277059] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 666.277059] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5255744e-c136-3a74-cba0-9f905829bbe0" [ 666.277059] env[65121]: _type = "Task" [ 666.277059] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.287605] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5255744e-c136-3a74-cba0-9f905829bbe0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.435297] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.435716] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.481371] env[65121]: DEBUG nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 666.506996] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 666.507264] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 666.507405] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 666.507573] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 666.507708] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 666.507847] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 666.508059] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.508210] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 666.508432] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 666.508614] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 666.508779] env[65121]: DEBUG nova.virt.hardware [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 666.509666] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4a850d-5871-48ba-bdb6-78b10de8dc4a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.519846] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8061fe0d-9725-4786-8e5f-8af6d6ebb69d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.549765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.127s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.552782] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.829s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.553653] env[65121]: INFO nova.compute.claims [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.565391] env[65121]: DEBUG oslo_vmware.api [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106086, 'name': RemoveSnapshot_Task, 'duration_secs': 0.681979} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.565391] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 666.566000] env[65121]: INFO nova.compute.manager [None req-a72ab8dc-76b9-4d14-a344-d1a76b1fcc94 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Took 17.21 seconds to snapshot the instance on the hypervisor. [ 666.573849] env[65121]: DEBUG nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 666.578603] env[65121]: INFO nova.scheduler.client.report [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted allocations for instance 5e915c8e-150a-4bfd-8005-a7a1a119057d [ 666.652346] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.652890] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.700969] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.704912] env[65121]: DEBUG oslo_vmware.api [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106087, 'name': PowerOnVM_Task, 'duration_secs': 0.677925} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.705238] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 666.705461] env[65121]: INFO nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Took 13.23 seconds to spawn the instance on the hypervisor. [ 666.706053] env[65121]: DEBUG nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 666.706552] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f5c4ae-6fbe-4588-a9db-77ee54374649 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.744449] env[65121]: DEBUG nova.compute.manager [req-44ec258a-0edb-490d-be2e-a57cab2be5ef req-c7473772-4bee-425d-8447-c1cfde728c3e service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-vif-plugged-3666dad5-ffa7-4c68-88b9-c2cb9d318615 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 666.744685] env[65121]: DEBUG oslo_concurrency.lockutils [req-44ec258a-0edb-490d-be2e-a57cab2be5ef req-c7473772-4bee-425d-8447-c1cfde728c3e service nova] Acquiring lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.744750] env[65121]: DEBUG oslo_concurrency.lockutils [req-44ec258a-0edb-490d-be2e-a57cab2be5ef req-c7473772-4bee-425d-8447-c1cfde728c3e service nova] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.744985] env[65121]: DEBUG oslo_concurrency.lockutils [req-44ec258a-0edb-490d-be2e-a57cab2be5ef req-c7473772-4bee-425d-8447-c1cfde728c3e service nova] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.745130] env[65121]: DEBUG nova.compute.manager [req-44ec258a-0edb-490d-be2e-a57cab2be5ef req-c7473772-4bee-425d-8447-c1cfde728c3e service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] No waiting events found dispatching network-vif-plugged-3666dad5-ffa7-4c68-88b9-c2cb9d318615 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 666.745310] env[65121]: WARNING nova.compute.manager [req-44ec258a-0edb-490d-be2e-a57cab2be5ef req-c7473772-4bee-425d-8447-c1cfde728c3e service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received unexpected event network-vif-plugged-3666dad5-ffa7-4c68-88b9-c2cb9d318615 for instance with vm_state building and task_state spawning. [ 666.758156] env[65121]: DEBUG oslo_concurrency.lockutils [req-def12d99-eeaf-477a-b961-5d037ad13090 req-6ec4a306-3e43-49c7-85b9-e07a19c0c1f5 service nova] Releasing lock "refresh_cache-7a0d0769-31c5-4ae1-8520-a744f64d39c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.787521] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5255744e-c136-3a74-cba0-9f905829bbe0, 'name': SearchDatastore_Task, 'duration_secs': 0.011322} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.787750] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.787997] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 61bd5995-701e-430f-9aae-4b266089e313/61bd5995-701e-430f-9aae-4b266089e313.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 666.788288] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da0ae843-a165-4b72-874d-0d12a4103444 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.797473] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 666.797473] env[65121]: value = "task-5106088" [ 666.797473] env[65121]: _type = "Task" [ 666.797473] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.805487] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.806570] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 666.837338] env[65121]: WARNING neutronclient.v2_0.client [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.838256] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.838863] env[65121]: WARNING openstack [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.005951] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.006561] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.105532] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1b70bd52-5ab8-431d-b615-2a5f16b947da tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "5e915c8e-150a-4bfd-8005-a7a1a119057d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.946s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.109397] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.230447] env[65121]: INFO nova.compute.manager [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Took 29.67 seconds to build instance. [ 667.267347] env[65121]: DEBUG nova.network.neutron [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Updating instance_info_cache with network_info: [{"id": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "address": "fa:16:3e:bb:cf:c0", "network": {"id": "e1cfb019-a6b4-4c9a-88d0-8cb6cdf5d0b8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1911550210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec5741daa3f3411bb74b3b965bbeef44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap813d0618-3f", "ovs_interfaceid": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.309042] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481602} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.310288] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.310983] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.311521] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.320219] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 61bd5995-701e-430f-9aae-4b266089e313/61bd5995-701e-430f-9aae-4b266089e313.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 667.320637] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 667.321496] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e96912c-2883-493e-a4b1-b5608f806a15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.329853] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 667.329853] env[65121]: value = "task-5106089" [ 667.329853] env[65121]: _type = "Task" [ 667.329853] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.343587] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106089, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.495620] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.495958] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.629849] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.630109] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.732939] env[65121]: DEBUG oslo_concurrency.lockutils [None req-911bfb2b-82e4-4710-b535-98bcedff3106 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "6218957b-6329-4004-97ca-07231b55937a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.198s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.735582] env[65121]: DEBUG nova.network.neutron [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Successfully updated port: f3966bcc-2651-436f-8bcb-da2f0b915c10 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 667.755803] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.756963] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.756963] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.769740] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Releasing lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.770235] env[65121]: DEBUG nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Instance network_info: |[{"id": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "address": "fa:16:3e:bb:cf:c0", "network": {"id": "e1cfb019-a6b4-4c9a-88d0-8cb6cdf5d0b8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1911550210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec5741daa3f3411bb74b3b965bbeef44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap813d0618-3f", "ovs_interfaceid": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 667.771147] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:cf:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32463b6d-4569-4755-8a29-873a028690a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '813d0618-3f84-4020-9e70-55c1bdcf5c22', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 667.778952] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Creating folder: Project (ec5741daa3f3411bb74b3b965bbeef44). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.780353] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f504e6e5-df2d-412b-94ef-8f3f24db281c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.794987] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Created folder: Project (ec5741daa3f3411bb74b3b965bbeef44) in parent group-v993268. [ 667.795261] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Creating folder: Instances. Parent ref: group-v993332. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 667.795530] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4dfb554-6653-44c8-a76d-73a5105772f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.808481] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Created folder: Instances in parent group-v993332. [ 667.808760] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 667.808961] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 667.809185] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6c5431cd-7004-463b-a589-77602dafb727 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.842755] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 667.842755] env[65121]: value = "task-5106092" [ 667.842755] env[65121]: _type = "Task" [ 667.842755] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.852826] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106089, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08257} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.856493] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 667.856493] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f6aa78-80e0-4407-9453-d12d8e1190a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.864357] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106092, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.886357] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] 61bd5995-701e-430f-9aae-4b266089e313/61bd5995-701e-430f-9aae-4b266089e313.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 667.889485] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b740be9-81f5-4448-947f-bd9dacca9452 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.913639] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 667.913639] env[65121]: value = "task-5106093" [ 667.913639] env[65121]: _type = "Task" [ 667.913639] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.925482] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106093, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.000391] env[65121]: DEBUG nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 668.008777] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.009619] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.090651] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ef2e49-b0dc-49ae-a305-27e89f4c762d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.100552] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812f4ed0-f833-419c-b955-93d089d75388 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.134468] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd635463-d65e-47c8-b957-5ba4cf8c94a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.145664] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51698a02-d168-4131-9a07-f657a789e1b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.162719] env[65121]: DEBUG nova.compute.provider_tree [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.229908] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.230327] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.241781] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "refresh_cache-c993d5a4-49d6-43aa-a1f0-4aac91fad953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.242075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired lock "refresh_cache-c993d5a4-49d6-43aa-a1f0-4aac91fad953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.242179] env[65121]: DEBUG nova.network.neutron [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 668.354804] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106092, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.374914] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Received event network-vif-plugged-a27cb922-5612-41b7-9637-f2c0bff9c8b5 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 668.375984] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquiring lock "6218957b-6329-4004-97ca-07231b55937a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.376687] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Lock "6218957b-6329-4004-97ca-07231b55937a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.376687] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Lock "6218957b-6329-4004-97ca-07231b55937a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.376687] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] No waiting events found dispatching network-vif-plugged-a27cb922-5612-41b7-9637-f2c0bff9c8b5 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 668.376866] env[65121]: WARNING nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Received unexpected event network-vif-plugged-a27cb922-5612-41b7-9637-f2c0bff9c8b5 for instance with vm_state active and task_state None. [ 668.376969] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Received event network-changed-a27cb922-5612-41b7-9637-f2c0bff9c8b5 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 668.377616] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Refreshing instance network info cache due to event network-changed-a27cb922-5612-41b7-9637-f2c0bff9c8b5. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 668.377616] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquiring lock "refresh_cache-6218957b-6329-4004-97ca-07231b55937a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.377616] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquired lock "refresh_cache-6218957b-6329-4004-97ca-07231b55937a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.378526] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Refreshing network info cache for port a27cb922-5612-41b7-9637-f2c0bff9c8b5 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 668.424936] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106093, 'name': ReconfigVM_Task, 'duration_secs': 0.305586} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.425252] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Reconfigured VM instance instance-00000016 to attach disk [datastore1] 61bd5995-701e-430f-9aae-4b266089e313/61bd5995-701e-430f-9aae-4b266089e313.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 668.425903] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-574e66b4-513c-44b1-9aa5-019e9bb1ea14 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.435222] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 668.435222] env[65121]: value = "task-5106094" [ 668.435222] env[65121]: _type = "Task" [ 668.435222] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.446329] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106094, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.514592] env[65121]: DEBUG nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 668.537782] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.642153] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.643365] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.644309] env[65121]: WARNING openstack [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.668130] env[65121]: DEBUG nova.scheduler.client.report [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 668.745261] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.748142] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.803197] env[65121]: DEBUG nova.network.neutron [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updating instance_info_cache with network_info: [{"id": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "address": "fa:16:3e:ae:e8:18", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3666dad5-ff", "ovs_interfaceid": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "142f7620-365e-4b15-8278-d8cee47cfccb", "address": "fa:16:3e:19:8b:7b", "network": {"id": "f63e0e62-428d-409d-ad1a-90381cb6bfe1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1252860029", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap142f7620-36", "ovs_interfaceid": "142f7620-365e-4b15-8278-d8cee47cfccb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "address": "fa:16:3e:fa:8e:7e", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aac0bc0-17", "ovs_interfaceid": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 668.859070] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106092, 'name': CreateVM_Task, 'duration_secs': 0.668683} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.859070] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 668.859070] env[65121]: WARNING neutronclient.v2_0.client [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.859070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.859373] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.859602] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 668.859853] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4dc26ea1-e48b-4d8a-9a12-c1d53b1148df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.865973] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 668.865973] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5252dd28-45a2-8abe-548a-e8e97504026b" [ 668.865973] env[65121]: _type = "Task" [ 668.865973] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.876785] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5252dd28-45a2-8abe-548a-e8e97504026b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.881625] env[65121]: WARNING neutronclient.v2_0.client [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.881625] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.881625] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.890405] env[65121]: DEBUG nova.network.neutron [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 668.945913] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106094, 'name': Rename_Task, 'duration_secs': 0.166171} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.946311] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 668.946651] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-238b698e-ba2f-4c16-ba17-c1cd210cb372 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.954186] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 668.954186] env[65121]: value = "task-5106095" [ 668.954186] env[65121]: _type = "Task" [ 668.954186] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.965941] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106095, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.022319] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.022726] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.051251] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.174130] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.174883] env[65121]: DEBUG nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 669.178848] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.129s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.179489] env[65121]: DEBUG nova.objects.instance [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 669.308626] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Releasing lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.309733] env[65121]: DEBUG nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Instance network_info: |[{"id": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "address": "fa:16:3e:ae:e8:18", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3666dad5-ff", "ovs_interfaceid": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "142f7620-365e-4b15-8278-d8cee47cfccb", "address": "fa:16:3e:19:8b:7b", "network": {"id": "f63e0e62-428d-409d-ad1a-90381cb6bfe1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1252860029", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap142f7620-36", "ovs_interfaceid": "142f7620-365e-4b15-8278-d8cee47cfccb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "address": "fa:16:3e:fa:8e:7e", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aac0bc0-17", "ovs_interfaceid": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 669.309985] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:e8:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9aa05ef8-c7bb-4af5-983f-bfa0f3f88223', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3666dad5-ffa7-4c68-88b9-c2cb9d318615', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:8b:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '142f7620-365e-4b15-8278-d8cee47cfccb', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:8e:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9aa05ef8-c7bb-4af5-983f-bfa0f3f88223', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 669.321938] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Creating folder: Project (bc44d2096bff489193193973fe3a3550). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 669.322661] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5db0e484-959c-4c4c-bf7d-03d2ea8cac7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.335591] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Created folder: Project (bc44d2096bff489193193973fe3a3550) in parent group-v993268. [ 669.335832] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Creating folder: Instances. Parent ref: group-v993335. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 669.341150] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d94699fe-9d8e-4261-bbfd-7a5fc7ee5d3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.353473] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Created folder: Instances in parent group-v993335. [ 669.353747] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 669.354168] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 669.354642] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7d1ac59-b110-4e8c-ab53-433c4fe2ddab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.392106] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5252dd28-45a2-8abe-548a-e8e97504026b, 'name': SearchDatastore_Task, 'duration_secs': 0.010848} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.394281] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.394651] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 669.395285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.395285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.395514] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 669.395905] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 669.395905] env[65121]: value = "task-5106098" [ 669.395905] env[65121]: _type = "Task" [ 669.395905] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.396233] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c1b0727-b361-4512-b950-ecbb2c8c52cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.410873] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106098, 'name': CreateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.414849] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 669.415182] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 669.416439] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ade0f193-7f2d-4a0b-9fdc-1dea7bba6fe1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.425456] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 669.425456] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ad81cf-01f2-510e-6eef-8cad588065a1" [ 669.425456] env[65121]: _type = "Task" [ 669.425456] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.440888] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ad81cf-01f2-510e-6eef-8cad588065a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.461330] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.461749] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.473486] env[65121]: DEBUG oslo_vmware.api [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106095, 'name': PowerOnVM_Task, 'duration_secs': 0.506684} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.474025] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 669.474305] env[65121]: INFO nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Took 9.84 seconds to spawn the instance on the hypervisor. [ 669.474607] env[65121]: DEBUG nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 669.475676] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43690ea-4a07-4f8f-9c2a-7480e5449b17 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.493274] env[65121]: WARNING neutronclient.v2_0.client [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.493961] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.494473] env[65121]: WARNING openstack [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.694504] env[65121]: DEBUG nova.compute.utils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 669.695517] env[65121]: DEBUG nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 669.695923] env[65121]: DEBUG nova.network.neutron [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 669.699863] env[65121]: WARNING neutronclient.v2_0.client [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.699863] env[65121]: WARNING neutronclient.v2_0.client [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.699863] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.699863] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.711602] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7914b742-51b8-46de-ad57-e575db10d24f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.722751] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Suspending the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 669.722751] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-d8f18e90-477f-4c4b-afa2-128e7c8f382f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.731030] env[65121]: DEBUG oslo_vmware.api [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] Waiting for the task: (returnval){ [ 669.731030] env[65121]: value = "task-5106099" [ 669.731030] env[65121]: _type = "Task" [ 669.731030] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.742051] env[65121]: DEBUG oslo_vmware.api [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] Task: {'id': task-5106099, 'name': SuspendVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.784729] env[65121]: WARNING neutronclient.v2_0.client [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.786675] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.787043] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.795462] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.795725] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.827126] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.827566] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.911794] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106098, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.922224] env[65121]: DEBUG nova.network.neutron [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Updating instance_info_cache with network_info: [{"id": "f3966bcc-2651-436f-8bcb-da2f0b915c10", "address": "fa:16:3e:b9:40:1d", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3966bcc-26", "ovs_interfaceid": "f3966bcc-2651-436f-8bcb-da2f0b915c10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 669.939499] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ad81cf-01f2-510e-6eef-8cad588065a1, 'name': SearchDatastore_Task, 'duration_secs': 0.027196} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.943949] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea7324fd-05be-4063-bcce-1d0e1aeeab63 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.952628] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 669.952628] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e301ac-7eea-ba2c-e4c1-4cdcaa92ed6d" [ 669.952628] env[65121]: _type = "Task" [ 669.952628] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.963785] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e301ac-7eea-ba2c-e4c1-4cdcaa92ed6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.998989] env[65121]: INFO nova.compute.manager [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Took 31.28 seconds to build instance. [ 670.197845] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03747cfb-98d3-458e-8bf1-02ee32dd868f tempest-ServersAdmin275Test-1143244090 tempest-ServersAdmin275Test-1143244090-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.198336] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.743s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.200066] env[65121]: INFO nova.compute.claims [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.207192] env[65121]: DEBUG nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 670.244656] env[65121]: DEBUG oslo_vmware.api [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] Task: {'id': task-5106099, 'name': SuspendVM_Task} progress is 45%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.253434] env[65121]: DEBUG nova.policy [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd443765da2574f558da7dbdc0528f89f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a92a7a59dd5c44f9b925030ae81e344d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 670.305264] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Updated VIF entry in instance network info cache for port a27cb922-5612-41b7-9637-f2c0bff9c8b5. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 670.305571] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Updating instance_info_cache with network_info: [{"id": "a27cb922-5612-41b7-9637-f2c0bff9c8b5", "address": "fa:16:3e:ba:1a:a2", "network": {"id": "e62fba86-0632-4afa-9490-1123f16bd48e", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-744677946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b816d77d6f3b4bc7a20d5b97ba068006", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4223acd2-30f7-440e-b975-60b30d931694", "external-id": "nsx-vlan-transportzone-647", "segmentation_id": 647, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa27cb922-56", "ovs_interfaceid": "a27cb922-5612-41b7-9637-f2c0bff9c8b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 670.416902] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106098, 'name': CreateVM_Task, 'duration_secs': 0.81912} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.417322] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 670.417935] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.418585] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.419249] env[65121]: WARNING neutronclient.v2_0.client [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.419735] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.420408] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.421097] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 670.421173] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acfb7216-033c-4120-8f16-a12907314ac1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.425333] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Releasing lock "refresh_cache-c993d5a4-49d6-43aa-a1f0-4aac91fad953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.425333] env[65121]: DEBUG nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Instance network_info: |[{"id": "f3966bcc-2651-436f-8bcb-da2f0b915c10", "address": "fa:16:3e:b9:40:1d", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3966bcc-26", "ovs_interfaceid": "f3966bcc-2651-436f-8bcb-da2f0b915c10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 670.427526] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:40:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3966bcc-2651-436f-8bcb-da2f0b915c10', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 670.439751] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Creating folder: Project (5724c072955c4cd3a5500dedf3b972a3). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.439751] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62b74234-c63e-4915-90f6-c140dec4a1ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.444836] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 670.444836] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f5e304-2349-6a66-c9c5-146887cfb13c" [ 670.444836] env[65121]: _type = "Task" [ 670.444836] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.456781] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f5e304-2349-6a66-c9c5-146887cfb13c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.462671] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Created folder: Project (5724c072955c4cd3a5500dedf3b972a3) in parent group-v993268. [ 670.462671] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Creating folder: Instances. Parent ref: group-v993338. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.464032] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-510a7633-d0f5-4f2e-a3c9-3f4aedc30d90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.471049] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e301ac-7eea-ba2c-e4c1-4cdcaa92ed6d, 'name': SearchDatastore_Task, 'duration_secs': 0.015256} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.471890] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.472336] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 4b6fe349-335e-4202-ab76-04f6ccc036ee/4b6fe349-335e-4202-ab76-04f6ccc036ee.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 670.472641] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77e29224-3a06-4515-bb97-461134862268 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.480109] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 670.480109] env[65121]: value = "task-5106102" [ 670.480109] env[65121]: _type = "Task" [ 670.480109] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.485523] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Created folder: Instances in parent group-v993338. [ 670.485860] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 670.486576] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 670.486851] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f0ebb13-ed0f-4c05-b6f0-9de93f92c544 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.507386] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b421eda-8816-4167-b4e3-647dd2410d6c tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.804s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.507714] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106102, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.514707] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 670.514707] env[65121]: value = "task-5106103" [ 670.514707] env[65121]: _type = "Task" [ 670.514707] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.524459] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106103, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.687434] env[65121]: DEBUG nova.network.neutron [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Successfully created port: 4f357400-7a42-4e3d-994c-a29afc1b366f {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 670.745415] env[65121]: DEBUG oslo_vmware.api [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] Task: {'id': task-5106099, 'name': SuspendVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.808113] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Releasing lock "refresh_cache-6218957b-6329-4004-97ca-07231b55937a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.809323] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Received event network-vif-plugged-1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 670.809566] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquiring lock "61bd5995-701e-430f-9aae-4b266089e313-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 670.809769] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Lock "61bd5995-701e-430f-9aae-4b266089e313-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.809932] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Lock "61bd5995-701e-430f-9aae-4b266089e313-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.810107] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] No waiting events found dispatching network-vif-plugged-1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 670.810260] env[65121]: WARNING nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Received unexpected event network-vif-plugged-1c577738-fcc1-4754-9cee-519ac385ae80 for instance with vm_state building and task_state spawning. [ 670.810817] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Received event network-changed-1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 670.810817] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Refreshing instance network info cache due to event network-changed-1c577738-fcc1-4754-9cee-519ac385ae80. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 670.810817] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquiring lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.811028] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquired lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.811028] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Refreshing network info cache for port 1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 670.958592] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f5e304-2349-6a66-c9c5-146887cfb13c, 'name': SearchDatastore_Task, 'duration_secs': 0.01459} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.958913] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.959275] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 670.959533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.959684] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.959866] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 670.960177] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68be6f12-b771-4e34-84ed-8b8991b029c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.977863] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 670.978089] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 670.978840] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-543482cb-b974-4e27-a228-27d9c6944cc7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.990646] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 670.990646] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52673196-5205-657b-d8c1-890faaaf5e48" [ 670.990646] env[65121]: _type = "Task" [ 670.990646] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.994357] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106102, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.008147] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52673196-5205-657b-d8c1-890faaaf5e48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.011341] env[65121]: DEBUG nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 671.031535] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106103, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.230802] env[65121]: DEBUG nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 671.243618] env[65121]: DEBUG oslo_vmware.api [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] Task: {'id': task-5106099, 'name': SuspendVM_Task, 'duration_secs': 1.065316} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.243970] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Suspended the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 671.244375] env[65121]: DEBUG nova.compute.manager [None req-aa7284f4-1969-485d-a5ab-728d7c56f91a tempest-ServersAdminNegativeTestJSON-269013314 tempest-ServersAdminNegativeTestJSON-269013314-project-admin] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 671.244979] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e4f5e0-8b6e-4c8e-963c-80e5ebeaf1a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.265255] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 671.265458] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 671.265610] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 671.265783] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 671.265922] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 671.266243] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 671.266517] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.266688] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 671.266857] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 671.267017] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 671.267184] env[65121]: DEBUG nova.virt.hardware [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 671.268651] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79917d40-3535-4665-b959-e9704fa808a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.278179] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a63f88a-19da-41c6-8558-963d1a320266 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.319596] env[65121]: WARNING neutronclient.v2_0.client [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.320314] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.320662] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.492217] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106102, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.830797} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.492546] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 4b6fe349-335e-4202-ab76-04f6ccc036ee/4b6fe349-335e-4202-ab76-04f6ccc036ee.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 671.492704] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 671.495446] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3aabe1e3-8eae-4308-ab68-e821e23e3395 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.507684] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52673196-5205-657b-d8c1-890faaaf5e48, 'name': SearchDatastore_Task, 'duration_secs': 0.06084} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.510362] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 671.510362] env[65121]: value = "task-5106104" [ 671.510362] env[65121]: _type = "Task" [ 671.510362] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.510362] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ecadeb8-4794-470c-839a-4ed3e57c085e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.533218] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 671.533218] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5222da20-13f0-86d8-88fe-055250d64670" [ 671.533218] env[65121]: _type = "Task" [ 671.533218] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.544052] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106104, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.544052] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106103, 'name': CreateVM_Task, 'duration_secs': 0.869873} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.544593] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 671.545151] env[65121]: WARNING neutronclient.v2_0.client [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.545664] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.545821] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.546202] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 671.546491] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f0f567-ec9e-485f-8cad-260d42c071cc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.552075] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5222da20-13f0-86d8-88fe-055250d64670, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.556532] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.559295] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 671.559295] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5207db2c-55c8-2885-e776-19a21941708d" [ 671.559295] env[65121]: _type = "Task" [ 671.559295] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.569255] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5207db2c-55c8-2885-e776-19a21941708d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.807615] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c064bc65-6ced-41f6-97da-2f386980f3f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.820545] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97abf58-708b-44f1-8708-0e5119a880e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.872429] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7dc1864-bb75-451c-aa8a-7054e979d1b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.884578] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6456db4d-3c33-46e6-8487-34b2ab48bcfa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.906685] env[65121]: DEBUG nova.compute.provider_tree [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.025145] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106104, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.244762} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.025465] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 672.026357] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99cc72e-ae0f-418b-a74c-dc2dd2de18f2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.051328] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] 4b6fe349-335e-4202-ab76-04f6ccc036ee/4b6fe349-335e-4202-ab76-04f6ccc036ee.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 672.055382] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35f27a68-8873-47e2-a81b-2addb7854475 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.080971] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5222da20-13f0-86d8-88fe-055250d64670, 'name': SearchDatastore_Task, 'duration_secs': 0.033299} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.081960] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.082334] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b9c78b3b-dd43-4199-98a7-ff57fd129f43/b9c78b3b-dd43-4199-98a7-ff57fd129f43.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 672.084208] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a986ff63-b7ba-49f9-9791-75f752001987 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.086366] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 672.086366] env[65121]: value = "task-5106105" [ 672.086366] env[65121]: _type = "Task" [ 672.086366] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.090737] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5207db2c-55c8-2885-e776-19a21941708d, 'name': SearchDatastore_Task, 'duration_secs': 0.033696} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.094640] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.094880] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 672.095250] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.095327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.095454] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 672.095739] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26e5a68e-e98e-4ce6-97a6-d48f9fb6f00a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.099501] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 672.099501] env[65121]: value = "task-5106106" [ 672.099501] env[65121]: _type = "Task" [ 672.099501] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.109214] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.110032] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 672.110312] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 672.111318] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41e843e6-ad7c-49dc-89de-99cb865083a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.118180] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "6218957b-6329-4004-97ca-07231b55937a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.118395] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "6218957b-6329-4004-97ca-07231b55937a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.118592] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "6218957b-6329-4004-97ca-07231b55937a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.118700] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "6218957b-6329-4004-97ca-07231b55937a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.118862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "6218957b-6329-4004-97ca-07231b55937a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.120544] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.121127] env[65121]: INFO nova.compute.manager [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Terminating instance [ 672.126890] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 672.126890] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d306d2-1b12-13ca-71b6-9c021a197c15" [ 672.126890] env[65121]: _type = "Task" [ 672.126890] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.136624] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d306d2-1b12-13ca-71b6-9c021a197c15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.218209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "6164d64c-75c5-469e-ac32-667842b3717a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.218209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "6164d64c-75c5-469e-ac32-667842b3717a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.249430] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.249699] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.410200] env[65121]: DEBUG nova.scheduler.client.report [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 672.454921] env[65121]: DEBUG nova.network.neutron [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Successfully updated port: 4f357400-7a42-4e3d-994c-a29afc1b366f {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 672.604919] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106105, 'name': ReconfigVM_Task, 'duration_secs': 0.368201} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.611785] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Reconfigured VM instance instance-00000017 to attach disk [datastore1] 4b6fe349-335e-4202-ab76-04f6ccc036ee/4b6fe349-335e-4202-ab76-04f6ccc036ee.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 672.612577] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2eaece83-d9c1-4833-9b12-7a8cf7f239c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.622323] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106106, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.623865] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 672.623865] env[65121]: value = "task-5106107" [ 672.623865] env[65121]: _type = "Task" [ 672.623865] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.628672] env[65121]: DEBUG nova.compute.manager [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 672.628672] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 672.629732] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36dbc8c-8d54-4c3b-9fe0-1a7883399e66 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.643289] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106107, 'name': Rename_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.650265] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 672.650629] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d306d2-1b12-13ca-71b6-9c021a197c15, 'name': SearchDatastore_Task, 'duration_secs': 0.015405} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.650760] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75d2f116-7269-47ef-a040-eb17b444459c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.653320] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6f9b6b9-633a-4995-8ef7-086bb5b6af79 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.663046] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 672.663046] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cd9cdb-cb4a-610d-6508-570e5c698fe1" [ 672.663046] env[65121]: _type = "Task" [ 672.663046] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.663781] env[65121]: DEBUG oslo_vmware.api [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 672.663781] env[65121]: value = "task-5106108" [ 672.663781] env[65121]: _type = "Task" [ 672.663781] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.677131] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cd9cdb-cb4a-610d-6508-570e5c698fe1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.680782] env[65121]: DEBUG oslo_vmware.api [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106108, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.918524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.720s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.919136] env[65121]: DEBUG nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 672.921831] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.242s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.922070] env[65121]: DEBUG nova.objects.instance [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lazy-loading 'resources' on Instance uuid 20396b81-2041-4bf2-85c4-8ea66c838ec4 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 672.961100] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "refresh_cache-124876fe-a4f2-4140-8dee-d8bb72b5c9a5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.961100] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired lock "refresh_cache-124876fe-a4f2-4140-8dee-d8bb72b5c9a5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.961100] env[65121]: DEBUG nova.network.neutron [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 673.069646] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.070084] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.115672] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609676} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.116023] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b9c78b3b-dd43-4199-98a7-ff57fd129f43/b9c78b3b-dd43-4199-98a7-ff57fd129f43.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 673.116348] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 673.116531] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02410865-3841-4746-bba9-acdad544d397 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.124212] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 673.124212] env[65121]: value = "task-5106109" [ 673.124212] env[65121]: _type = "Task" [ 673.124212] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.138883] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106107, 'name': Rename_Task, 'duration_secs': 0.199042} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.142469] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 673.142821] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106109, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.143059] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad5994da-53fb-4173-9c36-09f0b9c90ffe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.152748] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 673.152748] env[65121]: value = "task-5106110" [ 673.152748] env[65121]: _type = "Task" [ 673.152748] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.162222] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106110, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.182026] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cd9cdb-cb4a-610d-6508-570e5c698fe1, 'name': SearchDatastore_Task, 'duration_secs': 0.022697} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.187640] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.187932] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] c993d5a4-49d6-43aa-a1f0-4aac91fad953/c993d5a4-49d6-43aa-a1f0-4aac91fad953.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 673.188242] env[65121]: DEBUG oslo_vmware.api [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106108, 'name': PowerOffVM_Task, 'duration_secs': 0.269761} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.188446] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f09a3b4a-a439-4cec-b172-9c10b9575d24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.190903] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 673.191089] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 673.191362] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c31b86d-841a-4544-9457-7724631d2d89 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.202098] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 673.202098] env[65121]: value = "task-5106111" [ 673.202098] env[65121]: _type = "Task" [ 673.202098] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.216401] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106111, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.283565] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 673.283565] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 673.283565] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Deleting the datastore file [datastore1] 6218957b-6329-4004-97ca-07231b55937a {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 673.283565] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-baed673c-2968-4de1-81da-3085560f9869 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.292386] env[65121]: DEBUG oslo_vmware.api [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for the task: (returnval){ [ 673.292386] env[65121]: value = "task-5106113" [ 673.292386] env[65121]: _type = "Task" [ 673.292386] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.302118] env[65121]: DEBUG oslo_vmware.api [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.428158] env[65121]: DEBUG nova.compute.utils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 673.441148] env[65121]: DEBUG nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 673.441148] env[65121]: DEBUG nova.network.neutron [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 673.441148] env[65121]: WARNING neutronclient.v2_0.client [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.441148] env[65121]: WARNING neutronclient.v2_0.client [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.441464] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.443058] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.463055] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.463509] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.636337] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106109, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080785} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.640018] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 673.640018] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a239cd86-e402-4ba9-9408-05aa4c6b9b88 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.647274] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-changed-3666dad5-ffa7-4c68-88b9-c2cb9d318615 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 673.647274] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Refreshing instance network info cache due to event network-changed-3666dad5-ffa7-4c68-88b9-c2cb9d318615. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 673.647274] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.647274] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquired lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 673.647274] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Refreshing network info cache for port 3666dad5-ffa7-4c68-88b9-c2cb9d318615 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 673.682379] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] b9c78b3b-dd43-4199-98a7-ff57fd129f43/b9c78b3b-dd43-4199-98a7-ff57fd129f43.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 673.688138] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa41d2d8-4484-4a1b-9190-ed176623e528 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.717270] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106110, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.719185] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 673.719185] env[65121]: value = "task-5106114" [ 673.719185] env[65121]: _type = "Task" [ 673.719185] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.727113] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106111, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.736779] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106114, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.812321] env[65121]: DEBUG oslo_vmware.api [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.932553] env[65121]: DEBUG nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 674.102126] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167749a2-64cc-4278-b8b5-2363a9279e63 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.116795] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b2c82a-6ada-4b42-97c5-164d65861f33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.149775] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82eea9b5-9353-4eeb-b054-fcc52d2517cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.158289] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fe0f2d-dffb-47ca-9479-ffc9bf79c1fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.173325] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.174072] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.174458] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.181691] env[65121]: DEBUG nova.compute.provider_tree [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.193270] env[65121]: DEBUG oslo_vmware.api [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106110, 'name': PowerOnVM_Task, 'duration_secs': 0.916828} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.194259] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 674.194490] env[65121]: INFO nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Took 10.40 seconds to spawn the instance on the hypervisor. [ 674.194736] env[65121]: DEBUG nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 674.195499] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1045068b-14c6-4333-9500-34269b89acf0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.219323] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106111, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.923186} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.219708] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] c993d5a4-49d6-43aa-a1f0-4aac91fad953/c993d5a4-49d6-43aa-a1f0-4aac91fad953.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 674.219899] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 674.224389] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-95b6091a-4276-4d85-976e-648f3c6207c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.233769] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106114, 'name': ReconfigVM_Task, 'duration_secs': 0.408672} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.234482] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Reconfigured VM instance instance-00000015 to attach disk [datastore1] b9c78b3b-dd43-4199-98a7-ff57fd129f43/b9c78b3b-dd43-4199-98a7-ff57fd129f43.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 674.238024] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 674.238024] env[65121]: value = "task-5106115" [ 674.238024] env[65121]: _type = "Task" [ 674.238024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.238024] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c790f766-a98f-4e61-babc-8764ebd1d961 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.249241] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106115, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.249817] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 674.249817] env[65121]: value = "task-5106116" [ 674.249817] env[65121]: _type = "Task" [ 674.249817] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.260061] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106116, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.276415] env[65121]: DEBUG nova.network.neutron [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 674.305706] env[65121]: WARNING neutronclient.v2_0.client [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.306612] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.306960] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.318151] env[65121]: DEBUG oslo_vmware.api [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Task: {'id': task-5106113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.542362} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.318410] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 674.318588] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 674.318755] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 674.318923] env[65121]: INFO nova.compute.manager [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] [instance: 6218957b-6329-4004-97ca-07231b55937a] Took 1.69 seconds to destroy the instance on the hypervisor. [ 674.319198] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 674.319388] env[65121]: DEBUG nova.compute.manager [-] [instance: 6218957b-6329-4004-97ca-07231b55937a] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 674.319482] env[65121]: DEBUG nova.network.neutron [-] [instance: 6218957b-6329-4004-97ca-07231b55937a] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 674.319716] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.320250] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.320496] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.390210] env[65121]: DEBUG nova.policy [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c21a3c41d3d64ef4aff6ab64ad2c8d7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5724c072955c4cd3a5500dedf3b972a3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 674.481118] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.483230] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.574619] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.689388] env[65121]: DEBUG nova.scheduler.client.report [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 674.721990] env[65121]: INFO nova.compute.manager [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Took 33.10 seconds to build instance. [ 674.750423] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106115, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073994} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.750423] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 674.753486] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10df938d-3709-49d8-afad-c3fb01b815c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.796302] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] c993d5a4-49d6-43aa-a1f0-4aac91fad953/c993d5a4-49d6-43aa-a1f0-4aac91fad953.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 674.796302] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106116, 'name': Rename_Task, 'duration_secs': 0.216404} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.796749] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edd213e9-4c14-4bff-b841-d9a5aef35410 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.815550] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 674.820811] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-614f5033-9b17-4b6e-b36f-c90f1005978e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.829620] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 674.829620] env[65121]: value = "task-5106117" [ 674.829620] env[65121]: _type = "Task" [ 674.829620] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.834518] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 674.834518] env[65121]: value = "task-5106118" [ 674.834518] env[65121]: _type = "Task" [ 674.834518] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.849026] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updated VIF entry in instance network info cache for port 1c577738-fcc1-4754-9cee-519ac385ae80. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 674.849026] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updating instance_info_cache with network_info: [{"id": "1c577738-fcc1-4754-9cee-519ac385ae80", "address": "fa:16:3e:ed:26:a9", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c577738-fc", "ovs_interfaceid": "1c577738-fcc1-4754-9cee-519ac385ae80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 674.849026] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106117, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.863874] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106118, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.909088] env[65121]: DEBUG nova.network.neutron [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Successfully created port: a56899a4-844b-4ea3-b942-aa09fda98ec3 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 674.943582] env[65121]: DEBUG nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 674.950161] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.950772] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.978328] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 674.978512] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 674.978640] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 674.978921] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 674.979183] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 674.980372] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 674.980372] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.980372] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 674.980632] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 674.980768] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 674.981969] env[65121]: DEBUG nova.virt.hardware [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 674.982844] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564d0aa4-a73b-4271-8cc1-a1f8353bdeae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.994973] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68e101f-e07e-4dfe-896a-5ec9e2e18da1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.041280] env[65121]: WARNING neutronclient.v2_0.client [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.042380] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.042491] env[65121]: WARNING openstack [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.196471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.275s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.199226] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.200s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.224171] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f00abe0a-5b25-47b9-a027-acb6afbf8df6 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.512s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.228897] env[65121]: INFO nova.scheduler.client.report [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Deleted allocations for instance 20396b81-2041-4bf2-85c4-8ea66c838ec4 [ 675.347050] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106117, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.352142] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Releasing lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 675.352444] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Received event network-changed-5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 675.352627] env[65121]: DEBUG nova.compute.manager [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing instance network info cache due to event network-changed-5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 675.352842] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquiring lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.352986] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Acquired lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.353178] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 675.354501] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106118, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.610561] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.611282] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.611622] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.664197] env[65121]: DEBUG nova.network.neutron [-] [instance: 6218957b-6329-4004-97ca-07231b55937a] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 675.737699] env[65121]: DEBUG nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 675.747109] env[65121]: DEBUG nova.network.neutron [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Updating instance_info_cache with network_info: [{"id": "4f357400-7a42-4e3d-994c-a29afc1b366f", "address": "fa:16:3e:55:8b:62", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.25", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f357400-7a", "ovs_interfaceid": "4f357400-7a42-4e3d-994c-a29afc1b366f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 675.761494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-567afaa2-6031-4651-af18-12173ccd4ea0 tempest-ServersAdmin275Test-1363827893 tempest-ServersAdmin275Test-1363827893-project-member] Lock "20396b81-2041-4bf2-85c4-8ea66c838ec4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.703s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.846605] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106118, 'name': ReconfigVM_Task, 'duration_secs': 0.726105} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.850124] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Reconfigured VM instance instance-00000018 to attach disk [datastore1] c993d5a4-49d6-43aa-a1f0-4aac91fad953/c993d5a4-49d6-43aa-a1f0-4aac91fad953.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 675.850426] env[65121]: DEBUG oslo_vmware.api [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106117, 'name': PowerOnVM_Task, 'duration_secs': 0.571612} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.854833] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1612f7a8-2872-40f2-bd1a-e95539c2e0f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.858221] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 675.858562] env[65121]: INFO nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Took 18.97 seconds to spawn the instance on the hypervisor. [ 675.858661] env[65121]: DEBUG nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 675.859782] env[65121]: WARNING neutronclient.v2_0.client [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.859933] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.862292] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.868745] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17ee13d-9060-46fd-a744-cef735bf1235 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.884776] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 675.884776] env[65121]: value = "task-5106119" [ 675.884776] env[65121]: _type = "Task" [ 675.884776] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.166396] env[65121]: INFO nova.compute.manager [-] [instance: 6218957b-6329-4004-97ca-07231b55937a] Took 1.85 seconds to deallocate network for instance. [ 676.243650] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7e1fc4-cfb3-41cb-ad98-a7afab221d0f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.257977] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fc3e64-6178-4c1e-a60a-178c90ebc283 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.262495] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Releasing lock "refresh_cache-124876fe-a4f2-4140-8dee-d8bb72b5c9a5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.262495] env[65121]: DEBUG nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Instance network_info: |[{"id": "4f357400-7a42-4e3d-994c-a29afc1b366f", "address": "fa:16:3e:55:8b:62", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.25", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f357400-7a", "ovs_interfaceid": "4f357400-7a42-4e3d-994c-a29afc1b366f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 676.263126] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:8b:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f357400-7a42-4e3d-994c-a29afc1b366f', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 676.274209] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 676.303878] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 676.305022] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.305325] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d8d81d5-b841-49d0-aded-4913ac16d80a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.323621] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741cc21a-3e93-4643-967e-5e0300425f9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.334264] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60a369d-c661-4452-a510-1e12a3d40e3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.338571] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 676.338571] env[65121]: value = "task-5106120" [ 676.338571] env[65121]: _type = "Task" [ 676.338571] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.341713] env[65121]: DEBUG nova.compute.manager [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-vif-plugged-1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 676.341967] env[65121]: DEBUG oslo_concurrency.lockutils [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Acquiring lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.342228] env[65121]: DEBUG oslo_concurrency.lockutils [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.342371] env[65121]: DEBUG oslo_concurrency.lockutils [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.342533] env[65121]: DEBUG nova.compute.manager [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] No waiting events found dispatching network-vif-plugged-1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 676.342688] env[65121]: WARNING nova.compute.manager [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received unexpected event network-vif-plugged-1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e for instance with vm_state active and task_state None. [ 676.342835] env[65121]: DEBUG nova.compute.manager [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-changed-1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 676.342975] env[65121]: DEBUG nova.compute.manager [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Refreshing instance network info cache due to event network-changed-1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 676.343151] env[65121]: DEBUG oslo_concurrency.lockutils [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Acquiring lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.357207] env[65121]: DEBUG nova.compute.provider_tree [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.365497] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106120, 'name': CreateVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.398490] env[65121]: INFO nova.compute.manager [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Took 38.05 seconds to build instance. [ 676.404335] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106119, 'name': Rename_Task, 'duration_secs': 0.223569} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.404992] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 676.405119] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3db3b49-01bb-4ecf-be51-eacfbe953de9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.414238] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 676.414238] env[65121]: value = "task-5106121" [ 676.414238] env[65121]: _type = "Task" [ 676.414238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.426853] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106121, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.625544] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 676.625544] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 676.675293] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.770201] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "c1004635-b318-489c-9e16-6cb545279953" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.770537] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c1004635-b318-489c-9e16-6cb545279953" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.820135] env[65121]: DEBUG nova.network.neutron [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Successfully updated port: a56899a4-844b-4ea3-b942-aa09fda98ec3 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 676.852758] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106120, 'name': CreateVM_Task, 'duration_secs': 0.469027} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.852953] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 676.853502] env[65121]: WARNING neutronclient.v2_0.client [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.853855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.854065] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.854417] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 676.854732] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1afe1ea-bb36-436c-b98f-0323cfece87e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.861714] env[65121]: DEBUG nova.scheduler.client.report [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 676.870016] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 676.870016] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bec05c-6502-e0d2-6f26-be6d6b34086c" [ 676.870016] env[65121]: _type = "Task" [ 676.870016] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.873858] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updated VIF entry in instance network info cache for port 3666dad5-ffa7-4c68-88b9-c2cb9d318615. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 676.874476] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updating instance_info_cache with network_info: [{"id": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "address": "fa:16:3e:ae:e8:18", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3666dad5-ff", "ovs_interfaceid": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "142f7620-365e-4b15-8278-d8cee47cfccb", "address": "fa:16:3e:19:8b:7b", "network": {"id": "f63e0e62-428d-409d-ad1a-90381cb6bfe1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1252860029", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap142f7620-36", "ovs_interfaceid": "142f7620-365e-4b15-8278-d8cee47cfccb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "address": "fa:16:3e:fa:8e:7e", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aac0bc0-17", "ovs_interfaceid": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 676.888764] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bec05c-6502-e0d2-6f26-be6d6b34086c, 'name': SearchDatastore_Task, 'duration_secs': 0.021147} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.890039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.890039] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 676.890039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.891113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.891113] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 676.891484] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9adb5feb-1edc-4dcb-9bd4-c0876f8bc01a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.907647] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 676.907647] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 676.907647] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b8cce6f-2ade-4e4c-96a4-74d08361d9fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.909241] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a165d0b9-2b77-4b74-a346-b6d6f99fda19 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.574s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.914953] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 676.914953] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52193018-2841-f766-4cc1-df9332b16587" [ 676.914953] env[65121]: _type = "Task" [ 676.914953] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.915469] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.915690] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.915927] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.916272] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.916385] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.926616] env[65121]: INFO nova.compute.manager [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Terminating instance [ 676.935369] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52193018-2841-f766-4cc1-df9332b16587, 'name': SearchDatastore_Task, 'duration_secs': 0.013478} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.940659] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106121, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.945842] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcce7ee8-043b-421c-a99b-40b06e03045c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.953882] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 676.953882] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52844046-4f6f-f6a7-d6aa-e5d934ced1d8" [ 676.953882] env[65121]: _type = "Task" [ 676.953882] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.965116] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52844046-4f6f-f6a7-d6aa-e5d934ced1d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.035680] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.036204] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.137048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 677.137048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 677.137048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 677.137048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 677.137048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 677.137048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 677.137048] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 677.137048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 677.295761] env[65121]: WARNING neutronclient.v2_0.client [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.299087] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.299087] env[65121]: WARNING openstack [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.327308] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "refresh_cache-cd86f815-4b46-45b1-9a1c-ab2e28f63626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.327308] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired lock "refresh_cache-cd86f815-4b46-45b1-9a1c-ab2e28f63626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.327308] env[65121]: DEBUG nova.network.neutron [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 677.382720] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Releasing lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.382881] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-vif-plugged-142f7620-365e-4b15-8278-d8cee47cfccb {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 677.383133] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.384049] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.384049] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 677.384049] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] No waiting events found dispatching network-vif-plugged-142f7620-365e-4b15-8278-d8cee47cfccb {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 677.384049] env[65121]: WARNING nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received unexpected event network-vif-plugged-142f7620-365e-4b15-8278-d8cee47cfccb for instance with vm_state building and task_state spawning. [ 677.384049] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-changed-142f7620-365e-4b15-8278-d8cee47cfccb {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 677.384945] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Refreshing instance network info cache due to event network-changed-142f7620-365e-4b15-8278-d8cee47cfccb. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 677.384945] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.384945] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquired lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.384945] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Refreshing network info cache for port 142f7620-365e-4b15-8278-d8cee47cfccb {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 677.412714] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 677.429495] env[65121]: DEBUG oslo_vmware.api [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106121, 'name': PowerOnVM_Task, 'duration_secs': 0.601274} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.429495] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 677.429495] env[65121]: INFO nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Took 10.95 seconds to spawn the instance on the hypervisor. [ 677.429495] env[65121]: DEBUG nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 677.430304] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1dae82-d1ca-4e36-abe0-82fa8ae4ca9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.440375] env[65121]: DEBUG nova.compute.manager [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 677.440375] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 677.444039] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa00c213-bab3-4462-9db8-d11b890eefd3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.457682] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 677.463413] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-019fba76-ea0e-43ad-bfff-90b091496744 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.479689] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52844046-4f6f-f6a7-d6aa-e5d934ced1d8, 'name': SearchDatastore_Task, 'duration_secs': 0.013366} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.480454] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.480454] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 124876fe-a4f2-4140-8dee-d8bb72b5c9a5/124876fe-a4f2-4140-8dee-d8bb72b5c9a5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 677.482860] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7544558-7bae-498a-ba50-ba55cb95ed7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.490821] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 677.490821] env[65121]: value = "task-5106123" [ 677.490821] env[65121]: _type = "Task" [ 677.490821] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.504822] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.522355] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updated VIF entry in instance network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 677.524410] env[65121]: DEBUG nova.network.neutron [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 677.576910] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 677.577245] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 677.577494] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore2] 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 677.577924] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee8f1f8c-2d2d-4d1b-bb84-69c8429258db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.587979] env[65121]: DEBUG oslo_vmware.api [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 677.587979] env[65121]: value = "task-5106124" [ 677.587979] env[65121]: _type = "Task" [ 677.587979] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.598765] env[65121]: DEBUG oslo_vmware.api [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.640110] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.835077] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.835651] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.891331] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.692s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 677.895098] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.895817] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.896818] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.906323] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.719s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.907781] env[65121]: INFO nova.compute.claims [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.949398] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.961336] env[65121]: INFO nova.compute.manager [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Took 34.73 seconds to build instance. [ 678.006408] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106123, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.026945] env[65121]: DEBUG oslo_concurrency.lockutils [req-21b314ff-f279-4596-acbb-d6b832e0e9df req-9544a9f3-0866-4a67-ab32-cd98534a64a8 service nova] Releasing lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.103622] env[65121]: DEBUG oslo_vmware.api [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.163300] env[65121]: DEBUG nova.network.neutron [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 678.287325] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.287819] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.411622] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.411622] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.449793] env[65121]: WARNING neutronclient.v2_0.client [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.451419] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.451419] env[65121]: WARNING openstack [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.464916] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e804af6-c0fc-49e4-b989-a9af5e85db5a tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.561s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 678.505165] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106123, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699467} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.506778] env[65121]: INFO nova.scheduler.client.report [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Deleted allocation for migration 5a7d1565-69d9-4c80-b059-fdb739cebe0a [ 678.507978] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 124876fe-a4f2-4140-8dee-d8bb72b5c9a5/124876fe-a4f2-4140-8dee-d8bb72b5c9a5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 678.508607] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 678.511955] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d94151f4-e9eb-4926-a6ef-05dbb1a64d21 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.530567] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 678.530567] env[65121]: value = "task-5106125" [ 678.530567] env[65121]: _type = "Task" [ 678.530567] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.543668] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106125, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.603362] env[65121]: DEBUG oslo_vmware.api [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.551753} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.604877] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.605630] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.605976] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.615141] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 678.615251] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 678.615387] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.615515] env[65121]: INFO nova.compute.manager [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Took 1.18 seconds to destroy the instance on the hypervisor. [ 678.615831] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 678.616720] env[65121]: DEBUG nova.compute.manager [-] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 678.616774] env[65121]: DEBUG nova.network.neutron [-] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 678.617035] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.617494] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.617749] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.643455] env[65121]: DEBUG nova.network.neutron [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Updating instance_info_cache with network_info: [{"id": "a56899a4-844b-4ea3-b942-aa09fda98ec3", "address": "fa:16:3e:ad:12:77", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa56899a4-84", "ovs_interfaceid": "a56899a4-844b-4ea3-b942-aa09fda98ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 678.758877] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.836877] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updated VIF entry in instance network info cache for port 142f7620-365e-4b15-8278-d8cee47cfccb. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 678.837187] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updating instance_info_cache with network_info: [{"id": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "address": "fa:16:3e:ae:e8:18", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3666dad5-ff", "ovs_interfaceid": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "142f7620-365e-4b15-8278-d8cee47cfccb", "address": "fa:16:3e:19:8b:7b", "network": {"id": "f63e0e62-428d-409d-ad1a-90381cb6bfe1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1252860029", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap142f7620-36", "ovs_interfaceid": "142f7620-365e-4b15-8278-d8cee47cfccb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "address": "fa:16:3e:fa:8e:7e", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aac0bc0-17", "ovs_interfaceid": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 678.970934] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 679.024046] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3fd7663c-894c-4bed-a2bc-0ba9b1e7eaff tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 19.922s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.051103] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106125, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084602} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.051381] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 679.052242] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b62267-bbae-4599-b507-73c994885bcc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.079460] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] 124876fe-a4f2-4140-8dee-d8bb72b5c9a5/124876fe-a4f2-4140-8dee-d8bb72b5c9a5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 679.084779] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99c4ea86-3952-49cc-a0ae-29b4321db13e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.110892] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 679.110892] env[65121]: value = "task-5106126" [ 679.110892] env[65121]: _type = "Task" [ 679.110892] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.123738] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106126, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.147195] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Releasing lock "refresh_cache-cd86f815-4b46-45b1-9a1c-ab2e28f63626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.147565] env[65121]: DEBUG nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Instance network_info: |[{"id": "a56899a4-844b-4ea3-b942-aa09fda98ec3", "address": "fa:16:3e:ad:12:77", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa56899a4-84", "ovs_interfaceid": "a56899a4-844b-4ea3-b942-aa09fda98ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 679.148178] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:12:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a56899a4-844b-4ea3-b942-aa09fda98ec3', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 679.158035] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 679.158164] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 679.158361] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e52f368-248a-4a97-889e-81c15d8eec15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.184018] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 679.184018] env[65121]: value = "task-5106127" [ 679.184018] env[65121]: _type = "Task" [ 679.184018] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.194098] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106127, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.340306] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Releasing lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.340762] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Received event network-vif-plugged-813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 679.340895] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "4b6fe349-335e-4202-ab76-04f6ccc036ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 679.341020] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 679.341193] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.341351] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] No waiting events found dispatching network-vif-plugged-813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 679.341523] env[65121]: WARNING nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Received unexpected event network-vif-plugged-813d0618-3f84-4020-9e70-55c1bdcf5c22 for instance with vm_state building and task_state spawning. [ 679.341681] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Received event network-changed-813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 679.341879] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Refreshing instance network info cache due to event network-changed-813d0618-3f84-4020-9e70-55c1bdcf5c22. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 679.342050] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.342202] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquired lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.342361] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Refreshing network info cache for port 813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 679.343873] env[65121]: DEBUG oslo_concurrency.lockutils [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Acquired lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.344084] env[65121]: DEBUG nova.network.neutron [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Refreshing network info cache for port 1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 679.498220] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 679.564937] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4294eb7-3b6c-4dd8-9dd4-177bc183ad6a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.579057] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5b53ca-80b4-4db4-8487-69a180bae42e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.627476] env[65121]: DEBUG nova.network.neutron [-] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 679.633337] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9630c72b-8399-4cd3-ae7f-58913b1e1599 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.649148] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c51fd8f-9bd8-459f-afd0-4b549bd6e5ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.654473] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106126, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.671309] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 679.696366] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106127, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.851009] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 679.851883] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.852303] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.860328] env[65121]: WARNING neutronclient.v2_0.client [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 679.861013] env[65121]: WARNING openstack [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.861460] env[65121]: WARNING openstack [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.003883] env[65121]: DEBUG nova.compute.manager [req-a34ec478-5fd6-41ef-ba1c-93e9191c1794 req-24dd0f3d-ebe2-4d83-b93f-8f51e1edae52 service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Received event network-vif-plugged-a56899a4-844b-4ea3-b942-aa09fda98ec3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.006378] env[65121]: DEBUG oslo_concurrency.lockutils [req-a34ec478-5fd6-41ef-ba1c-93e9191c1794 req-24dd0f3d-ebe2-4d83-b93f-8f51e1edae52 service nova] Acquiring lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.006378] env[65121]: DEBUG oslo_concurrency.lockutils [req-a34ec478-5fd6-41ef-ba1c-93e9191c1794 req-24dd0f3d-ebe2-4d83-b93f-8f51e1edae52 service nova] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.006378] env[65121]: DEBUG oslo_concurrency.lockutils [req-a34ec478-5fd6-41ef-ba1c-93e9191c1794 req-24dd0f3d-ebe2-4d83-b93f-8f51e1edae52 service nova] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.006378] env[65121]: DEBUG nova.compute.manager [req-a34ec478-5fd6-41ef-ba1c-93e9191c1794 req-24dd0f3d-ebe2-4d83-b93f-8f51e1edae52 service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] No waiting events found dispatching network-vif-plugged-a56899a4-844b-4ea3-b942-aa09fda98ec3 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 680.006378] env[65121]: WARNING nova.compute.manager [req-a34ec478-5fd6-41ef-ba1c-93e9191c1794 req-24dd0f3d-ebe2-4d83-b93f-8f51e1edae52 service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Received unexpected event network-vif-plugged-a56899a4-844b-4ea3-b942-aa09fda98ec3 for instance with vm_state building and task_state spawning. [ 680.043030] env[65121]: WARNING openstack [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.043281] env[65121]: WARNING openstack [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.068680] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.070422] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.111566] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "06a99a15-5c8a-4be0-b393-80a104b9ba76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.111790] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.131793] env[65121]: INFO nova.compute.manager [-] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Took 1.51 seconds to deallocate network for instance. [ 680.158478] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106126, 'name': ReconfigVM_Task, 'duration_secs': 0.70972} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.158903] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Reconfigured VM instance instance-00000019 to attach disk [datastore2] 124876fe-a4f2-4140-8dee-d8bb72b5c9a5/124876fe-a4f2-4140-8dee-d8bb72b5c9a5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.159956] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca4c70b7-b48e-4a57-91d6-408694d5cc93 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.168229] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 680.168229] env[65121]: value = "task-5106128" [ 680.168229] env[65121]: _type = "Task" [ 680.168229] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.181519] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106128, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.187275] env[65121]: WARNING neutronclient.v2_0.client [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.189386] env[65121]: WARNING openstack [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.189386] env[65121]: WARNING openstack [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.211711] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106127, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.218203] env[65121]: ERROR nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [req-982a36c5-6c77-41d2-be8a-50feebd05044] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-982a36c5-6c77-41d2-be8a-50feebd05044"}]} [ 680.245380] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 680.248935] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.249986] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.250649] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.271651] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 680.273145] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 680.292189] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 680.314837] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 680.408390] env[65121]: DEBUG nova.network.neutron [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updated VIF entry in instance network info cache for port 1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 680.409053] env[65121]: DEBUG nova.network.neutron [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updating instance_info_cache with network_info: [{"id": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "address": "fa:16:3e:ae:e8:18", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3666dad5-ff", "ovs_interfaceid": "3666dad5-ffa7-4c68-88b9-c2cb9d318615", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "142f7620-365e-4b15-8278-d8cee47cfccb", "address": "fa:16:3e:19:8b:7b", "network": {"id": "f63e0e62-428d-409d-ad1a-90381cb6bfe1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1252860029", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap142f7620-36", "ovs_interfaceid": "142f7620-365e-4b15-8278-d8cee47cfccb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "address": "fa:16:3e:fa:8e:7e", "network": {"id": "b471ea5e-5fb6-46dc-b07a-defab14faaca", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1164787234", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9aa05ef8-c7bb-4af5-983f-bfa0f3f88223", "external-id": "nsx-vlan-transportzone-135", "segmentation_id": 135, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aac0bc0-17", "ovs_interfaceid": "1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.483416] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Updated VIF entry in instance network info cache for port 813d0618-3f84-4020-9e70-55c1bdcf5c22. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 680.483416] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Updating instance_info_cache with network_info: [{"id": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "address": "fa:16:3e:bb:cf:c0", "network": {"id": "e1cfb019-a6b4-4c9a-88d0-8cb6cdf5d0b8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1911550210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec5741daa3f3411bb74b3b965bbeef44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap813d0618-3f", "ovs_interfaceid": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.654762] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.681992] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106128, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.713033] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106127, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.728182] env[65121]: DEBUG nova.compute.manager [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Received event network-vif-plugged-4f357400-7a42-4e3d-994c-a29afc1b366f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.728182] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Acquiring lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.728400] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.728550] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.728768] env[65121]: DEBUG nova.compute.manager [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] No waiting events found dispatching network-vif-plugged-4f357400-7a42-4e3d-994c-a29afc1b366f {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 680.728944] env[65121]: WARNING nova.compute.manager [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Received unexpected event network-vif-plugged-4f357400-7a42-4e3d-994c-a29afc1b366f for instance with vm_state building and task_state spawning. [ 680.729107] env[65121]: DEBUG nova.compute.manager [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Received event network-changed-4f357400-7a42-4e3d-994c-a29afc1b366f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.729279] env[65121]: DEBUG nova.compute.manager [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Refreshing instance network info cache due to event network-changed-4f357400-7a42-4e3d-994c-a29afc1b366f. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 680.729495] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Acquiring lock "refresh_cache-124876fe-a4f2-4140-8dee-d8bb72b5c9a5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.729562] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Acquired lock "refresh_cache-124876fe-a4f2-4140-8dee-d8bb72b5c9a5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.729672] env[65121]: DEBUG nova.network.neutron [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Refreshing network info cache for port 4f357400-7a42-4e3d-994c-a29afc1b366f {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 680.820036] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d3b268-75dc-4eab-808c-8f27e1492ba5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.829800] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86ddfc1-a016-49ba-8ffc-85f039a8b3c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.863119] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7dbf67-3ef2-4f33-a963-33fee2e971c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.872394] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099d01f4-a31c-4a68-a4c8-bbd059be71d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.888416] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 680.912113] env[65121]: DEBUG oslo_concurrency.lockutils [req-88ebe959-7264-4456-a5f8-bdee9d7d6f71 req-a1b6bb61-137e-4913-b9c4-7e6293741a85 service nova] Releasing lock "refresh_cache-b9c78b3b-dd43-4199-98a7-ff57fd129f43" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.987859] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Releasing lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.988252] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Received event network-vif-plugged-f3966bcc-2651-436f-8bcb-da2f0b915c10 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.988464] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.988715] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.988944] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.989148] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] No waiting events found dispatching network-vif-plugged-f3966bcc-2651-436f-8bcb-da2f0b915c10 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 680.989316] env[65121]: WARNING nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Received unexpected event network-vif-plugged-f3966bcc-2651-436f-8bcb-da2f0b915c10 for instance with vm_state building and task_state spawning. [ 680.993050] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Received event network-changed-f3966bcc-2651-436f-8bcb-da2f0b915c10 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.993050] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Refreshing instance network info cache due to event network-changed-f3966bcc-2651-436f-8bcb-da2f0b915c10. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 680.993050] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "refresh_cache-c993d5a4-49d6-43aa-a1f0-4aac91fad953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.993050] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquired lock "refresh_cache-c993d5a4-49d6-43aa-a1f0-4aac91fad953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.993050] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Refreshing network info cache for port f3966bcc-2651-436f-8bcb-da2f0b915c10 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 681.063276] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.064152] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.188823] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106128, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.212606] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106127, 'name': CreateVM_Task, 'duration_secs': 1.650721} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.212796] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 681.213331] env[65121]: WARNING neutronclient.v2_0.client [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.213724] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.213879] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 681.214400] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 681.214685] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a6a227a-c3d8-4ec3-a002-1014c812ad07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.220575] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 681.220575] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52498147-11eb-0293-e0a2-6418aa100109" [ 681.220575] env[65121]: _type = "Task" [ 681.220575] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.230585] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52498147-11eb-0293-e0a2-6418aa100109, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.232394] env[65121]: WARNING neutronclient.v2_0.client [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.233130] env[65121]: WARNING openstack [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.233657] env[65121]: WARNING openstack [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.357016] env[65121]: WARNING openstack [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.357016] env[65121]: WARNING openstack [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.414658] env[65121]: ERROR nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [req-b41d4361-f40f-46dd-97b9-fa1dad22ad99] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b41d4361-f40f-46dd-97b9-fa1dad22ad99"}]} [ 681.425605] env[65121]: WARNING neutronclient.v2_0.client [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.426311] env[65121]: WARNING openstack [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.426631] env[65121]: WARNING openstack [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.437095] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 681.475742] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 681.475742] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 681.495064] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: 4a33a8e0-77cd-44c4-8f32-409d761653c9 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 681.497297] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.498762] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.499389] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.527202] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 681.605037] env[65121]: DEBUG nova.network.neutron [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Updated VIF entry in instance network info cache for port 4f357400-7a42-4e3d-994c-a29afc1b366f. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 681.605037] env[65121]: DEBUG nova.network.neutron [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Updating instance_info_cache with network_info: [{"id": "4f357400-7a42-4e3d-994c-a29afc1b366f", "address": "fa:16:3e:55:8b:62", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.25", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f357400-7a", "ovs_interfaceid": "4f357400-7a42-4e3d-994c-a29afc1b366f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 681.701339] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106128, 'name': Rename_Task, 'duration_secs': 1.319273} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.701339] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 681.701339] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e985162c-841f-41f7-b60d-3622f742d380 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.711126] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 681.711126] env[65121]: value = "task-5106129" [ 681.711126] env[65121]: _type = "Task" [ 681.711126] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.728315] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.747326] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52498147-11eb-0293-e0a2-6418aa100109, 'name': SearchDatastore_Task, 'duration_secs': 0.024106} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.747875] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.748283] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 681.748647] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.748890] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 681.749242] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 681.749644] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27bf7e30-16a4-402a-91c6-0857fdaef212 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.766403] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 681.766819] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 681.767368] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34ad5934-c01a-4422-b759-4460e4dc6f84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.775245] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.775674] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.784078] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 681.784078] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522ca0f5-8756-9712-1f3c-df4e2f8e8fbe" [ 681.784078] env[65121]: _type = "Task" [ 681.784078] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.799884] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522ca0f5-8756-9712-1f3c-df4e2f8e8fbe, 'name': SearchDatastore_Task} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.800763] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27a9cd69-12ef-49fa-9a3b-b33f10884372 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.809449] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 681.809449] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5284ed1f-b234-27ec-daeb-ac570bbb6013" [ 681.809449] env[65121]: _type = "Task" [ 681.809449] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.818904] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5284ed1f-b234-27ec-daeb-ac570bbb6013, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.062733] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "3f69040e-7df3-4535-a3a7-90f3348ef346" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.063227] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.063227] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "50083ee5-9655-4cab-9d50-04a97baac626" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.063346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "50083ee5-9655-4cab-9d50-04a97baac626" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.107388] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] Releasing lock "refresh_cache-124876fe-a4f2-4140-8dee-d8bb72b5c9a5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 682.107618] env[65121]: DEBUG nova.compute.manager [req-5f7e12fc-cc34-4bef-b0f7-9a689adf9ea3 req-77581fc3-badd-4a3f-8107-c0b7e8c75760 service nova] [instance: 6218957b-6329-4004-97ca-07231b55937a] Received event network-vif-deleted-a27cb922-5612-41b7-9637-f2c0bff9c8b5 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 682.145616] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460ca433-6982-454a-89a5-145e0923523d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.153618] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c1c306-5a7a-4082-9c0f-e377435c4682 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.187763] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11728ad-4c26-40f5-bc57-2696c2de1096 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.197524] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc2f8ef-9fa8-4546-87fb-7f6738c539d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.213790] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 682.226198] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106129, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.320928] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5284ed1f-b234-27ec-daeb-ac570bbb6013, 'name': SearchDatastore_Task, 'duration_secs': 0.024982} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.321215] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 682.322479] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] cd86f815-4b46-45b1-9a1c-ab2e28f63626/cd86f815-4b46-45b1-9a1c-ab2e28f63626.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 682.322479] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a56fdf9-0598-4cbb-8e50-49c829d89a02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.329338] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 682.329338] env[65121]: value = "task-5106130" [ 682.329338] env[65121]: _type = "Task" [ 682.329338] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.338371] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.561653] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 682.562386] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.562716] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.696603] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Updated VIF entry in instance network info cache for port f3966bcc-2651-436f-8bcb-da2f0b915c10. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 682.696975] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Updating instance_info_cache with network_info: [{"id": "f3966bcc-2651-436f-8bcb-da2f0b915c10", "address": "fa:16:3e:b9:40:1d", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3966bcc-26", "ovs_interfaceid": "f3966bcc-2651-436f-8bcb-da2f0b915c10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 682.725615] env[65121]: DEBUG oslo_vmware.api [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106129, 'name': PowerOnVM_Task, 'duration_secs': 0.720291} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.726826] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 682.726826] env[65121]: INFO nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Took 11.50 seconds to spawn the instance on the hypervisor. [ 682.726826] env[65121]: DEBUG nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 682.727476] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e864b31c-bc4c-4cab-a526-c9b77aaa86d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.760029] env[65121]: ERROR nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [req-22545205-8745-4867-9fa2-b723abbf6236] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-22545205-8745-4867-9fa2-b723abbf6236"}]} [ 682.783244] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 682.809591] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 682.809798] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 682.823618] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: 4a33a8e0-77cd-44c4-8f32-409d761653c9 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 682.840037] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.850301] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 683.202109] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Releasing lock "refresh_cache-c993d5a4-49d6-43aa-a1f0-4aac91fad953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 683.202109] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Received event network-changed-b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 683.202109] env[65121]: DEBUG nova.compute.manager [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Refreshing instance network info cache due to event network-changed-b25aa25b-2d1b-414e-883a-324894d4b483. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 683.202109] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquiring lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.202109] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Acquired lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.202599] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Refreshing network info cache for port b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 683.259996] env[65121]: INFO nova.compute.manager [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Took 37.57 seconds to build instance. [ 683.342716] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106130, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.497301] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee4eb73-99cb-44b5-8520-2eae0d4b931f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.510044] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91d9d5b-63ea-48b6-8417-adcbc47a57ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.566619] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44a4eb9-c7a8-44aa-8d18-a9e61e43fe13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.573608] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b477c3d-247a-424f-adc8-8b40a17768e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.589331] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 683.702274] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.702558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.702763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.703041] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.703390] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.707050] env[65121]: INFO nova.compute.manager [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Terminating instance [ 683.709023] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.709293] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.710174] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.763569] env[65121]: DEBUG oslo_concurrency.lockutils [None req-962e331a-6405-4880-8422-b3a3c4be3982 tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.361s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.840802] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106130, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.071931} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.842257] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] cd86f815-4b46-45b1-9a1c-ab2e28f63626/cd86f815-4b46-45b1-9a1c-ab2e28f63626.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 683.842483] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 683.844284] env[65121]: DEBUG nova.compute.manager [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Received event network-changed-a56899a4-844b-4ea3-b942-aa09fda98ec3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 683.844850] env[65121]: DEBUG nova.compute.manager [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Refreshing instance network info cache due to event network-changed-a56899a4-844b-4ea3-b942-aa09fda98ec3. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 683.845126] env[65121]: DEBUG oslo_concurrency.lockutils [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Acquiring lock "refresh_cache-cd86f815-4b46-45b1-9a1c-ab2e28f63626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.845311] env[65121]: DEBUG oslo_concurrency.lockutils [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Acquired lock "refresh_cache-cd86f815-4b46-45b1-9a1c-ab2e28f63626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.845476] env[65121]: DEBUG nova.network.neutron [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Refreshing network info cache for port a56899a4-844b-4ea3-b942-aa09fda98ec3 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 683.846724] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8667b7e-c839-49f9-b05b-dac22f13e48c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.856066] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 683.856066] env[65121]: value = "task-5106131" [ 683.856066] env[65121]: _type = "Task" [ 683.856066] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.870095] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106131, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.882604] env[65121]: DEBUG nova.compute.manager [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Received event network-vif-deleted-7aef5809-0a17-4d3e-b553-93f337e59e89 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 683.882858] env[65121]: DEBUG nova.compute.manager [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Received event network-changed-813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 683.883023] env[65121]: DEBUG nova.compute.manager [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Refreshing instance network info cache due to event network-changed-813d0618-3f84-4020-9e70-55c1bdcf5c22. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 683.883412] env[65121]: DEBUG oslo_concurrency.lockutils [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Acquiring lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.883412] env[65121]: DEBUG oslo_concurrency.lockutils [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Acquired lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.883713] env[65121]: DEBUG nova.network.neutron [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Refreshing network info cache for port 813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 684.140543] env[65121]: DEBUG nova.scheduler.client.report [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 48 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 684.140543] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 48 to 49 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 684.141171] env[65121]: DEBUG nova.compute.provider_tree [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 684.219888] env[65121]: DEBUG nova.compute.manager [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 684.220299] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.222322] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab92889f-b23d-48b6-8e7d-be20e9ff3cb0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.230233] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 684.230500] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a4dbf58-6abc-4208-bb7c-f05216866361 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.237320] env[65121]: DEBUG oslo_vmware.api [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 684.237320] env[65121]: value = "task-5106132" [ 684.237320] env[65121]: _type = "Task" [ 684.237320] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.238764] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.239208] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.263190] env[65121]: DEBUG oslo_vmware.api [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.266956] env[65121]: DEBUG nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 684.350568] env[65121]: WARNING neutronclient.v2_0.client [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 684.351375] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.352952] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.377102] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106131, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083929} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.377102] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 684.377589] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd850c3-eb2e-4f64-b77a-302c1b978f46 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.396175] env[65121]: WARNING neutronclient.v2_0.client [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 684.396836] env[65121]: WARNING openstack [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.397191] env[65121]: WARNING openstack [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.413599] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] cd86f815-4b46-45b1-9a1c-ab2e28f63626/cd86f815-4b46-45b1-9a1c-ab2e28f63626.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 684.415532] env[65121]: WARNING neutronclient.v2_0.client [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 684.416178] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.416677] env[65121]: WARNING openstack [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.425182] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27a7bb84-ab7e-4085-9cac-088e3bd6547a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.448331] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 684.448331] env[65121]: value = "task-5106133" [ 684.448331] env[65121]: _type = "Task" [ 684.448331] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.459387] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.646696] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 6.741s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.647268] env[65121]: DEBUG nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 684.649919] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.949s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.650150] env[65121]: DEBUG nova.objects.instance [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lazy-loading 'resources' on Instance uuid 6887bb66-7186-4d58-be59-d0c53c3221c6 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 684.700446] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Updated VIF entry in instance network info cache for port b25aa25b-2d1b-414e-883a-324894d4b483. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 684.700902] env[65121]: DEBUG nova.network.neutron [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Updating instance_info_cache with network_info: [{"id": "b25aa25b-2d1b-414e-883a-324894d4b483", "address": "fa:16:3e:0e:be:f6", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb25aa25b-2d", "ovs_interfaceid": "b25aa25b-2d1b-414e-883a-324894d4b483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 684.756060] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.756493] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.769182] env[65121]: DEBUG oslo_vmware.api [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106132, 'name': PowerOffVM_Task, 'duration_secs': 0.4393} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.769441] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 684.769596] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 684.769845] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1d43d12-648f-43d9-a8e3-a0490a731030 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.787735] env[65121]: WARNING openstack [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.788137] env[65121]: WARNING openstack [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.798727] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.944416] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 684.944747] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 684.944893] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Deleting the datastore file [datastore1] b9c78b3b-dd43-4199-98a7-ff57fd129f43 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 684.945315] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f30e2e9f-c899-4e3a-b3d1-5e91a9d5c41e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.954650] env[65121]: DEBUG oslo_vmware.api [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 684.954650] env[65121]: value = "task-5106135" [ 684.954650] env[65121]: _type = "Task" [ 684.954650] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.962985] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106133, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.972647] env[65121]: DEBUG oslo_vmware.api [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106135, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.056088] env[65121]: WARNING neutronclient.v2_0.client [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.056759] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.057125] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.117126] env[65121]: WARNING neutronclient.v2_0.client [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.117126] env[65121]: WARNING openstack [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.118083] env[65121]: WARNING openstack [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.163022] env[65121]: DEBUG nova.compute.utils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 685.166232] env[65121]: DEBUG nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 685.166232] env[65121]: DEBUG nova.network.neutron [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 685.166232] env[65121]: WARNING neutronclient.v2_0.client [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.166845] env[65121]: WARNING neutronclient.v2_0.client [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.167975] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.168898] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.204109] env[65121]: DEBUG oslo_concurrency.lockutils [req-911ed9c2-ebd9-43d6-b7df-bf6ce8b316fb req-732869b5-9e21-418e-a19a-e1aa29df756f service nova] Releasing lock "refresh_cache-e2d32a5e-c350-4b2b-9243-c3b412193a82" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.253995] env[65121]: DEBUG nova.policy [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff853d2df89c4bc18e74cb1dd62a789f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c0dbd078f6b42d5ad8eeb0a45cac9f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 685.289517] env[65121]: DEBUG nova.network.neutron [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Updated VIF entry in instance network info cache for port a56899a4-844b-4ea3-b942-aa09fda98ec3. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 685.289893] env[65121]: DEBUG nova.network.neutron [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Updating instance_info_cache with network_info: [{"id": "a56899a4-844b-4ea3-b942-aa09fda98ec3", "address": "fa:16:3e:ad:12:77", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa56899a4-84", "ovs_interfaceid": "a56899a4-844b-4ea3-b942-aa09fda98ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 685.323808] env[65121]: DEBUG nova.network.neutron [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Updated VIF entry in instance network info cache for port 813d0618-3f84-4020-9e70-55c1bdcf5c22. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 685.324794] env[65121]: DEBUG nova.network.neutron [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Updating instance_info_cache with network_info: [{"id": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "address": "fa:16:3e:bb:cf:c0", "network": {"id": "e1cfb019-a6b4-4c9a-88d0-8cb6cdf5d0b8", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1911550210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec5741daa3f3411bb74b3b965bbeef44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap813d0618-3f", "ovs_interfaceid": "813d0618-3f84-4020-9e70-55c1bdcf5c22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 685.467380] env[65121]: DEBUG oslo_vmware.api [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106135, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.225772} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.474837] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 685.474837] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 685.474837] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.474837] env[65121]: INFO nova.compute.manager [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Took 1.25 seconds to destroy the instance on the hypervisor. [ 685.475147] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 685.475234] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106133, 'name': ReconfigVM_Task, 'duration_secs': 0.545756} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.475696] env[65121]: DEBUG nova.compute.manager [-] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 685.475799] env[65121]: DEBUG nova.network.neutron [-] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 685.476066] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.476905] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.476966] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.485788] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Reconfigured VM instance instance-0000001a to attach disk [datastore2] cd86f815-4b46-45b1-9a1c-ab2e28f63626/cd86f815-4b46-45b1-9a1c-ab2e28f63626.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 685.486636] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-783720e3-2966-447e-9d38-19263f9615fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.498753] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 685.498753] env[65121]: value = "task-5106136" [ 685.498753] env[65121]: _type = "Task" [ 685.498753] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.514201] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106136, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.572814] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.659517] env[65121]: DEBUG nova.network.neutron [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Successfully created port: ea99cbd9-05be-40d9-952c-9bee73993946 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 685.682706] env[65121]: DEBUG nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 685.767359] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9ced9f-accc-458d-bf1e-0404cf458c88 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.777412] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5736a95a-b529-4cc7-b405-6e8d99df7656 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.811814] env[65121]: DEBUG oslo_concurrency.lockutils [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Releasing lock "refresh_cache-cd86f815-4b46-45b1-9a1c-ab2e28f63626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.811814] env[65121]: DEBUG nova.compute.manager [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 685.811814] env[65121]: DEBUG nova.compute.manager [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing instance network info cache due to event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 685.811814] env[65121]: DEBUG oslo_concurrency.lockutils [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Acquiring lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.812189] env[65121]: DEBUG oslo_concurrency.lockutils [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Acquired lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 685.812189] env[65121]: DEBUG nova.network.neutron [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 685.815375] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cd7ec9-5539-4d1c-a873-a9588d03b720 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.825676] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be635a4-539d-4d52-bd55-f099660a2ee7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.842045] env[65121]: DEBUG oslo_concurrency.lockutils [req-0badf90d-9be4-409b-a981-00dee80c25a8 req-12530b01-9104-4d2b-b767-fcdfd2751987 service nova] Releasing lock "refresh_cache-4b6fe349-335e-4202-ab76-04f6ccc036ee" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.842939] env[65121]: DEBUG nova.compute.provider_tree [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.009627] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106136, 'name': Rename_Task, 'duration_secs': 0.167053} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.009895] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 686.010152] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be3c7e4b-24ab-4dee-9362-c1ef8155dd3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.016704] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 686.016704] env[65121]: value = "task-5106137" [ 686.016704] env[65121]: _type = "Task" [ 686.016704] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.025041] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106137, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.315952] env[65121]: WARNING neutronclient.v2_0.client [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.316663] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.316950] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.346210] env[65121]: DEBUG nova.scheduler.client.report [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 686.529410] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106137, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.693509] env[65121]: DEBUG nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 686.722450] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 686.722690] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 686.722839] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 686.723082] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 686.723261] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 686.723407] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 686.723613] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.723764] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 686.723918] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 686.724158] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 686.724362] env[65121]: DEBUG nova.virt.hardware [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 686.725348] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d31114-c960-4cde-9a03-f548d03095a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.734540] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6afcc9d-e360-40f5-83ec-6df1cda82bf9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.852056] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.202s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 686.854249] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.746s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.855824] env[65121]: INFO nova.compute.claims [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.886411] env[65121]: INFO nova.scheduler.client.report [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Deleted allocations for instance 6887bb66-7186-4d58-be59-d0c53c3221c6 [ 686.944284] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.945279] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.035548] env[65121]: DEBUG oslo_vmware.api [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106137, 'name': PowerOnVM_Task, 'duration_secs': 0.921976} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.035840] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 687.036744] env[65121]: INFO nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Took 12.09 seconds to spawn the instance on the hypervisor. [ 687.036744] env[65121]: DEBUG nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 687.037610] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47614ca4-b507-410c-8109-00c40d03b7bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.047957] env[65121]: DEBUG nova.network.neutron [-] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 687.071592] env[65121]: WARNING neutronclient.v2_0.client [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.074077] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.074077] env[65121]: WARNING openstack [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.217243] env[65121]: DEBUG nova.network.neutron [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updated VIF entry in instance network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 687.217515] env[65121]: DEBUG nova.network.neutron [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updating instance_info_cache with network_info: [{"id": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "address": "fa:16:3e:47:bb:ea", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119a70b-b1", "ovs_interfaceid": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 687.276789] env[65121]: DEBUG nova.network.neutron [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Successfully updated port: ea99cbd9-05be-40d9-952c-9bee73993946 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 687.394778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f33415f0-9c42-4e11-873f-2507fc339943 tempest-ServerDiagnosticsV248Test-1593113287 tempest-ServerDiagnosticsV248Test-1593113287-project-member] Lock "6887bb66-7186-4d58-be59-d0c53c3221c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.178s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 687.552544] env[65121]: INFO nova.compute.manager [-] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Took 2.08 seconds to deallocate network for instance. [ 687.561492] env[65121]: INFO nova.compute.manager [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Took 32.21 seconds to build instance. [ 687.725345] env[65121]: DEBUG oslo_concurrency.lockutils [req-a87aad7d-e852-4b15-b1b3-a9059336ffa5 req-c534ff88-6558-4c74-a759-2d402a2c4f2e service nova] Releasing lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 687.784951] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.785380] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 687.785669] env[65121]: DEBUG nova.network.neutron [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 688.066929] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 688.067567] env[65121]: DEBUG oslo_concurrency.lockutils [None req-290134ef-dbfe-4363-b063-889549ad6c5f tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.899s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.289937] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.290449] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.521665] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc70e2bc-aed7-433b-9c37-c14c0ed79914 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.530729] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da77653a-b0f7-491a-abcd-b6bd3f07aaf1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.563895] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe9e9fe-e041-47ee-9937-fca73d36ea2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.572632] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a50cb32-e4fc-4b11-97e0-7b009983d911 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.578751] env[65121]: DEBUG nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 688.594392] env[65121]: DEBUG nova.compute.provider_tree [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.815402] env[65121]: DEBUG nova.compute.manager [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 688.816436] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fee1f1-396c-4e1e-87c1-86e7a76371cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.098506] env[65121]: DEBUG nova.scheduler.client.report [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.106279] env[65121]: DEBUG nova.network.neutron [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 689.115827] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.198177] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.198177] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.252768] env[65121]: DEBUG nova.compute.manager [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 689.253063] env[65121]: DEBUG nova.compute.manager [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing instance network info cache due to event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 689.253305] env[65121]: DEBUG oslo_concurrency.lockutils [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Acquiring lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.253447] env[65121]: DEBUG oslo_concurrency.lockutils [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Acquired lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.253817] env[65121]: DEBUG nova.network.neutron [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 689.330199] env[65121]: INFO nova.compute.manager [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] instance snapshotting [ 689.335103] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6308bd8b-099c-450d-92a2-7ed4735a3576 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.358709] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94256b3d-2fb9-460e-ac5d-b46749f9d47d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.610554] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.756s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.611150] env[65121]: DEBUG nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 689.614199] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.076s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.615733] env[65121]: INFO nova.compute.claims [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.719465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.719747] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.719966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.720186] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.720365] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.723038] env[65121]: INFO nova.compute.manager [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Terminating instance [ 689.759247] env[65121]: WARNING neutronclient.v2_0.client [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.759965] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.760298] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.871011] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 689.871339] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d50c60c8-633b-4933-b2e1-fbb27e4dc82e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.878819] env[65121]: WARNING neutronclient.v2_0.client [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.879742] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.880105] env[65121]: WARNING openstack [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.891646] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 689.891646] env[65121]: value = "task-5106138" [ 689.891646] env[65121]: _type = "Task" [ 689.891646] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.900418] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106138, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.121295] env[65121]: DEBUG nova.compute.utils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 690.124650] env[65121]: DEBUG nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 690.124650] env[65121]: DEBUG nova.network.neutron [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 690.125651] env[65121]: WARNING neutronclient.v2_0.client [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.125651] env[65121]: WARNING neutronclient.v2_0.client [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.127438] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.127438] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.227811] env[65121]: DEBUG nova.compute.manager [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 690.228062] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.228953] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e6e4d5-e779-44a3-bf7b-1012e155fbc9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.242299] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 690.242299] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3fd45f93-e2ce-43a2-88bd-a7c7618f706e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.252229] env[65121]: DEBUG oslo_vmware.api [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 690.252229] env[65121]: value = "task-5106139" [ 690.252229] env[65121]: _type = "Task" [ 690.252229] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.260355] env[65121]: DEBUG nova.policy [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc3c22580f4442dcae4a2b5245ede996', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '083c4f2aa7bf4167b472db6bebb9d8bd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 690.273887] env[65121]: DEBUG nova.network.neutron [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Updating instance_info_cache with network_info: [{"id": "ea99cbd9-05be-40d9-952c-9bee73993946", "address": "fa:16:3e:c1:43:51", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea99cbd9-05", "ovs_interfaceid": "ea99cbd9-05be-40d9-952c-9bee73993946", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 690.275376] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.275775] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.290047] env[65121]: DEBUG oslo_vmware.api [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.290949] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 690.291287] env[65121]: DEBUG nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Instance network_info: |[{"id": "ea99cbd9-05be-40d9-952c-9bee73993946", "address": "fa:16:3e:c1:43:51", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea99cbd9-05", "ovs_interfaceid": "ea99cbd9-05be-40d9-952c-9bee73993946", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 690.292493] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:43:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea99cbd9-05be-40d9-952c-9bee73993946', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 690.304027] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Creating folder: Project (3c0dbd078f6b42d5ad8eeb0a45cac9f5). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.304027] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63932e94-709f-4261-8fa3-70a43af106dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.317633] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Created folder: Project (3c0dbd078f6b42d5ad8eeb0a45cac9f5) in parent group-v993268. [ 690.317878] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Creating folder: Instances. Parent ref: group-v993343. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.323512] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7fb57082-486f-4fd7-955f-c8431f708cd1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.341367] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Created folder: Instances in parent group-v993343. [ 690.342142] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 690.342142] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 690.342142] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce01ee68-eecb-44c6-a8ad-ffb5589749f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.368276] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 690.368276] env[65121]: value = "task-5106142" [ 690.368276] env[65121]: _type = "Task" [ 690.368276] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.381699] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106142, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.384187] env[65121]: DEBUG nova.compute.manager [req-c5573a6e-0aa3-4e1f-8866-25075c0d6711 req-ee01db0f-2b81-4e9e-97d3-a508311cdf23 service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Received event network-vif-plugged-ea99cbd9-05be-40d9-952c-9bee73993946 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 690.384489] env[65121]: DEBUG oslo_concurrency.lockutils [req-c5573a6e-0aa3-4e1f-8866-25075c0d6711 req-ee01db0f-2b81-4e9e-97d3-a508311cdf23 service nova] Acquiring lock "c2f9fa49-4273-4885-b59d-0635ed739d65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.384818] env[65121]: DEBUG oslo_concurrency.lockutils [req-c5573a6e-0aa3-4e1f-8866-25075c0d6711 req-ee01db0f-2b81-4e9e-97d3-a508311cdf23 service nova] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.385038] env[65121]: DEBUG oslo_concurrency.lockutils [req-c5573a6e-0aa3-4e1f-8866-25075c0d6711 req-ee01db0f-2b81-4e9e-97d3-a508311cdf23 service nova] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.385336] env[65121]: DEBUG nova.compute.manager [req-c5573a6e-0aa3-4e1f-8866-25075c0d6711 req-ee01db0f-2b81-4e9e-97d3-a508311cdf23 service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] No waiting events found dispatching network-vif-plugged-ea99cbd9-05be-40d9-952c-9bee73993946 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 690.385615] env[65121]: WARNING nova.compute.manager [req-c5573a6e-0aa3-4e1f-8866-25075c0d6711 req-ee01db0f-2b81-4e9e-97d3-a508311cdf23 service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Received unexpected event network-vif-plugged-ea99cbd9-05be-40d9-952c-9bee73993946 for instance with vm_state building and task_state spawning. [ 690.408360] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106138, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.425043] env[65121]: WARNING neutronclient.v2_0.client [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.425872] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.426411] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.624909] env[65121]: DEBUG nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 690.706600] env[65121]: DEBUG nova.network.neutron [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updated VIF entry in instance network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 690.706600] env[65121]: DEBUG nova.network.neutron [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updating instance_info_cache with network_info: [{"id": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "address": "fa:16:3e:47:bb:ea", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119a70b-b1", "ovs_interfaceid": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 690.714025] env[65121]: DEBUG nova.network.neutron [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Successfully created port: c2770914-538c-4cb2-9133-0998f766b7ad {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 690.778356] env[65121]: DEBUG oslo_vmware.api [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106139, 'name': PowerOffVM_Task, 'duration_secs': 0.220708} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.778664] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 690.778819] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 690.779469] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcdc5cb9-2fb9-424a-952d-eaae10aef456 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.845401] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 690.845733] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 690.845999] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Deleting the datastore file [datastore2] 124876fe-a4f2-4140-8dee-d8bb72b5c9a5 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 690.846351] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34171dd5-ad95-4f18-adf5-f32612256d62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.855853] env[65121]: DEBUG oslo_vmware.api [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for the task: (returnval){ [ 690.855853] env[65121]: value = "task-5106144" [ 690.855853] env[65121]: _type = "Task" [ 690.855853] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.870687] env[65121]: DEBUG oslo_vmware.api [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106144, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.882017] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106142, 'name': CreateVM_Task, 'duration_secs': 0.413354} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.884657] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 690.885540] env[65121]: WARNING neutronclient.v2_0.client [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.886009] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.886980] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.887329] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 690.887592] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b5ee5a8-1eea-4a87-a850-9cb37a0dfe5d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.895080] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 690.895080] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fc91b3-778d-c190-4f6f-792a0ee9200b" [ 690.895080] env[65121]: _type = "Task" [ 690.895080] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.912997] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fc91b3-778d-c190-4f6f-792a0ee9200b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.917025] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106138, 'name': CreateSnapshot_Task, 'duration_secs': 0.749763} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.917025] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 690.917025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8218f921-74de-4743-8fe4-65cdef8be10e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.217389] env[65121]: DEBUG oslo_concurrency.lockutils [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Releasing lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.217389] env[65121]: DEBUG nova.compute.manager [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Received event network-changed-1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 691.217389] env[65121]: DEBUG nova.compute.manager [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Refreshing instance network info cache due to event network-changed-1c577738-fcc1-4754-9cee-519ac385ae80. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 691.217389] env[65121]: DEBUG oslo_concurrency.lockutils [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Acquiring lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.217389] env[65121]: DEBUG oslo_concurrency.lockutils [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Acquired lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.217389] env[65121]: DEBUG nova.network.neutron [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Refreshing network info cache for port 1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 691.370143] env[65121]: DEBUG oslo_vmware.api [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Task: {'id': task-5106144, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258218} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.370400] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 691.370934] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 691.370934] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.370934] env[65121]: INFO nova.compute.manager [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 691.371146] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 691.371330] env[65121]: DEBUG nova.compute.manager [-] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 691.371489] env[65121]: DEBUG nova.network.neutron [-] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 691.371755] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 691.372308] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 691.372567] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 691.380905] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd3fe97-508a-4d69-8411-509b50bd28cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.389212] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7376220-2676-4e9d-a11e-498d58a79d0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.427122] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ad9aa2-333a-449e-869c-a2ffaa715923 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.437257] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 691.443033] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1ac9c5d4-b913-426a-b2f1-2a433b6f47e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.446415] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fc91b3-778d-c190-4f6f-792a0ee9200b, 'name': SearchDatastore_Task, 'duration_secs': 0.030817} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.446657] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.451820] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.451820] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.451820] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.451820] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 691.451820] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2cac2c-b8ca-4640-9990-fb78afb60bbd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.454232] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4fd4341-2fa7-4c56-968f-ff4ab279fd52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.459754] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 691.459754] env[65121]: value = "task-5106145" [ 691.459754] env[65121]: _type = "Task" [ 691.459754] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.469159] env[65121]: DEBUG nova.compute.provider_tree [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.479493] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 691.479493] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 691.479493] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a963425c-da8a-4d17-9bee-4f23f0c63f50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.482219] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 691.482219] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524ddfe7-7a4b-77ae-227f-cea37a2fc40d" [ 691.482219] env[65121]: _type = "Task" [ 691.482219] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.485424] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106145, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.499191] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524ddfe7-7a4b-77ae-227f-cea37a2fc40d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.643266] env[65121]: DEBUG nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 691.674078] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 691.674078] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 691.674078] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 691.674078] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 691.674078] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 691.674078] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 691.674078] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.674462] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 691.674546] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 691.674685] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 691.674853] env[65121]: DEBUG nova.virt.hardware [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 691.675806] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f45518e-d65d-49fa-a5d6-c0da88a48cc5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.686116] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb6b94a-b36a-45b8-a391-1d3e325430d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.723754] env[65121]: WARNING neutronclient.v2_0.client [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 691.724464] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 691.724802] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 691.733402] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 691.794015] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "61bd5995-701e-430f-9aae-4b266089e313" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 691.794015] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 691.794015] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "61bd5995-701e-430f-9aae-4b266089e313-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 691.794015] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 691.794015] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.795102] env[65121]: INFO nova.compute.manager [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Terminating instance [ 691.979053] env[65121]: DEBUG nova.scheduler.client.report [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.987188] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106145, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.998139] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524ddfe7-7a4b-77ae-227f-cea37a2fc40d, 'name': SearchDatastore_Task, 'duration_secs': 0.023443} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.998980] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b212e5f8-2aa4-4133-8e3a-471e3e87bca9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.006872] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 692.006872] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524c607b-1213-febb-1b85-6d6b1b492040" [ 692.006872] env[65121]: _type = "Task" [ 692.006872] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.016096] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524c607b-1213-febb-1b85-6d6b1b492040, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.304625] env[65121]: DEBUG nova.compute.manager [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 692.305649] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 692.307057] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108e2743-ca32-4b0c-8384-41d65858ee90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.320920] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 692.321337] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e0e681f-e99d-401b-bb05-f5b89906216d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.330325] env[65121]: DEBUG oslo_vmware.api [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 692.330325] env[65121]: value = "task-5106146" [ 692.330325] env[65121]: _type = "Task" [ 692.330325] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.342353] env[65121]: DEBUG oslo_vmware.api [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.345360] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.345517] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.345580] env[65121]: DEBUG nova.compute.manager [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 692.346452] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d935f696-1a62-4989-bc11-e68c914eba0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.354908] env[65121]: DEBUG nova.compute.manager [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 692.355544] env[65121]: DEBUG nova.objects.instance [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'flavor' on Instance uuid c4096314-270a-4270-9e1d-5ace8ddbd286 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 692.479409] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106145, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.489558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.875s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.490114] env[65121]: DEBUG nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 692.494680] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.444s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.498487] env[65121]: INFO nova.compute.claims [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.519249] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524c607b-1213-febb-1b85-6d6b1b492040, 'name': SearchDatastore_Task, 'duration_secs': 0.014065} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.519769] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.519853] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c2f9fa49-4273-4885-b59d-0635ed739d65/c2f9fa49-4273-4885-b59d-0635ed739d65.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 692.520197] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-95428562-d452-4bb0-bb06-58b9a86ceb74 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.529464] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 692.529464] env[65121]: value = "task-5106147" [ 692.529464] env[65121]: _type = "Task" [ 692.529464] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.530602] env[65121]: DEBUG nova.network.neutron [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Successfully updated port: c2770914-538c-4cb2-9133-0998f766b7ad {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 692.542360] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106147, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.727656] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.728367] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.814803] env[65121]: WARNING neutronclient.v2_0.client [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.815281] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.815626] env[65121]: WARNING openstack [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.849318] env[65121]: DEBUG oslo_vmware.api [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106146, 'name': PowerOffVM_Task, 'duration_secs': 0.469836} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.849318] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 692.849459] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 692.851025] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39f8ffc1-0161-4949-ad66-ec14bfb49960 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.924383] env[65121]: DEBUG nova.network.neutron [-] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 692.940988] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 692.941395] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 692.941655] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Deleting the datastore file [datastore1] 61bd5995-701e-430f-9aae-4b266089e313 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 692.942567] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d365633-70c4-42ef-a8bc-5e2d81b8464c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.953443] env[65121]: DEBUG oslo_vmware.api [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 692.953443] env[65121]: value = "task-5106149" [ 692.953443] env[65121]: _type = "Task" [ 692.953443] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.965045] env[65121]: DEBUG oslo_vmware.api [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.985387] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106145, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.003443] env[65121]: DEBUG nova.compute.utils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 693.007674] env[65121]: DEBUG nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 693.007917] env[65121]: DEBUG nova.network.neutron [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 693.008270] env[65121]: WARNING neutronclient.v2_0.client [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.008642] env[65121]: WARNING neutronclient.v2_0.client [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.009504] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.009714] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.034467] env[65121]: DEBUG nova.network.neutron [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updated VIF entry in instance network info cache for port 1c577738-fcc1-4754-9cee-519ac385ae80. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 693.034910] env[65121]: DEBUG nova.network.neutron [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updating instance_info_cache with network_info: [{"id": "1c577738-fcc1-4754-9cee-519ac385ae80", "address": "fa:16:3e:ed:26:a9", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c577738-fc", "ovs_interfaceid": "1c577738-fcc1-4754-9cee-519ac385ae80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 693.040949] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "refresh_cache-27940143-16b5-4263-b23c-354ed8ea8866" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.040949] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "refresh_cache-27940143-16b5-4263-b23c-354ed8ea8866" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 693.040949] env[65121]: DEBUG nova.network.neutron [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 693.049896] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106147, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.150795] env[65121]: DEBUG nova.policy [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63bf1d576d7e44c3a8f99a75546efe52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4467da3ed41245ddbc93fc865a8b7bdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 693.371025] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 693.371025] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5683f71-c266-453b-b90e-ec5fed74c33c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.380022] env[65121]: DEBUG oslo_vmware.api [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 693.380022] env[65121]: value = "task-5106150" [ 693.380022] env[65121]: _type = "Task" [ 693.380022] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.389328] env[65121]: DEBUG oslo_vmware.api [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106150, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.428907] env[65121]: INFO nova.compute.manager [-] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Took 2.06 seconds to deallocate network for instance. [ 693.464245] env[65121]: DEBUG oslo_vmware.api [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106149, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.423352} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.464552] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 693.464761] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 693.465375] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.465375] env[65121]: INFO nova.compute.manager [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Took 1.16 seconds to destroy the instance on the hypervisor. [ 693.465467] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 693.465720] env[65121]: DEBUG nova.compute.manager [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 693.465822] env[65121]: DEBUG nova.network.neutron [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 693.466279] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.467736] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.468174] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.486868] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106145, 'name': CloneVM_Task, 'duration_secs': 1.996825} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.487231] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Created linked-clone VM from snapshot [ 693.488346] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1134f38-3652-400a-a9f4-94168282e7b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.497669] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Uploading image 45fa7707-1688-4fcb-aa2e-c4460117f6c3 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 693.508316] env[65121]: DEBUG nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 693.529497] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 693.529497] env[65121]: value = "vm-993347" [ 693.529497] env[65121]: _type = "VirtualMachine" [ 693.529497] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 693.529886] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f5583a89-c88b-4b31-ba2f-c36f960e307c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.538997] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lease: (returnval){ [ 693.538997] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d72986-adac-3214-204a-013c0b559fe5" [ 693.538997] env[65121]: _type = "HttpNfcLease" [ 693.538997] env[65121]: } obtained for exporting VM: (result){ [ 693.538997] env[65121]: value = "vm-993347" [ 693.538997] env[65121]: _type = "VirtualMachine" [ 693.538997] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 693.539406] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the lease: (returnval){ [ 693.539406] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d72986-adac-3214-204a-013c0b559fe5" [ 693.539406] env[65121]: _type = "HttpNfcLease" [ 693.539406] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 693.544243] env[65121]: DEBUG oslo_concurrency.lockutils [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] Releasing lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 693.544440] env[65121]: DEBUG nova.compute.manager [req-27ec151a-3d3d-4ba7-ba2f-a5693bd27591 req-5e2de894-39ef-4f23-858d-9bb0b5957d78 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-vif-deleted-142f7620-365e-4b15-8278-d8cee47cfccb {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 693.546145] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.546145] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.568986] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106147, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652187} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.571296] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c2f9fa49-4273-4885-b59d-0635ed739d65/c2f9fa49-4273-4885-b59d-0635ed739d65.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 693.571419] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.571691] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 693.571691] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d72986-adac-3214-204a-013c0b559fe5" [ 693.571691] env[65121]: _type = "HttpNfcLease" [ 693.571691] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 693.572131] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8ba5747-aad9-4efd-a33c-9051a0ae4965 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.574237] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 693.574237] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d72986-adac-3214-204a-013c0b559fe5" [ 693.574237] env[65121]: _type = "HttpNfcLease" [ 693.574237] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 693.575095] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a33884-a52a-42a9-9512-150b2c5d0a0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.587851] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c9dd22-6a3b-ef58-6eb7-989eea130a19/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 693.589013] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c9dd22-6a3b-ef58-6eb7-989eea130a19/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 693.590867] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 693.590867] env[65121]: value = "task-5106152" [ 693.590867] env[65121]: _type = "Task" [ 693.590867] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.665777] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106152, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.799896] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d2b4832c-7289-4d75-b2a5-ac71303766b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.836028] env[65121]: DEBUG nova.network.neutron [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Successfully created port: 0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 693.890319] env[65121]: DEBUG oslo_vmware.api [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106150, 'name': PowerOffVM_Task, 'duration_secs': 0.256903} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.891236] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.892891] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 693.893107] env[65121]: DEBUG nova.compute.manager [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 693.894322] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d09257-97f9-4e21-9f2f-5fe7d4b92062 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.934796] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.964682] env[65121]: DEBUG nova.network.neutron [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 694.068905] env[65121]: DEBUG nova.compute.manager [req-9f10810d-5f6c-41ee-b066-55b54553788d req-b12614ff-9631-48f4-be41-af27c74aeaa4 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-vif-deleted-1aac0bc0-17b0-4359-ba5c-2e5b2cc4037e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 694.068905] env[65121]: DEBUG nova.compute.manager [req-9f10810d-5f6c-41ee-b066-55b54553788d req-b12614ff-9631-48f4-be41-af27c74aeaa4 service nova] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Received event network-vif-deleted-3666dad5-ffa7-4c68-88b9-c2cb9d318615 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 694.106321] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106152, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127461} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.107227] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 694.108893] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649663d0-4cb6-4758-a1ab-d4017bc03a70 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.139688] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] c2f9fa49-4273-4885-b59d-0635ed739d65/c2f9fa49-4273-4885-b59d-0635ed739d65.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 694.143845] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-372e58f5-f54e-4c05-80a3-3e6249841d93 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.169875] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 694.169875] env[65121]: value = "task-5106153" [ 694.169875] env[65121]: _type = "Task" [ 694.169875] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.179942] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106153, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.223135] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82eb698f-048f-41ae-8e5e-4bd53c34c6d3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.233333] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a85599b-3360-4d6d-80ab-7ad52d27ad0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.272944] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723078ca-ee41-4bf1-bb3e-17e1f1642a56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.285796] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15aaed22-fa94-48de-b105-9f0c6432a25b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.303377] env[65121]: DEBUG nova.compute.provider_tree [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.409628] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df7315ea-7a30-422c-be55-e8a327039ac9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.064s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.518334] env[65121]: DEBUG nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 694.549486] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 694.549802] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 694.550104] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 694.550872] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 694.550872] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 694.551382] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 694.551855] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.551855] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 694.553022] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 694.553022] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 694.553022] env[65121]: DEBUG nova.virt.hardware [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 694.557921] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fa48b3-ebf9-4146-8b27-24fc9c910bc2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.568388] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bba7d8-ed79-419e-80e3-02d299c9ad9c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.682735] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106153, 'name': ReconfigVM_Task, 'duration_secs': 0.488314} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.683164] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Reconfigured VM instance instance-0000001b to attach disk [datastore2] c2f9fa49-4273-4885-b59d-0635ed739d65/c2f9fa49-4273-4885-b59d-0635ed739d65.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 694.683949] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccb81908-811d-4270-9e98-d368e3ddbf16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.691756] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 694.691756] env[65121]: value = "task-5106154" [ 694.691756] env[65121]: _type = "Task" [ 694.691756] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.705035] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106154, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.728861] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 694.729029] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.829704] env[65121]: ERROR nova.scheduler.client.report [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [req-3e45f9d9-ab4e-4422-a760-d975f5d47917] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3e45f9d9-ab4e-4422-a760-d975f5d47917"}]} [ 694.854047] env[65121]: DEBUG nova.scheduler.client.report [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 694.878947] env[65121]: DEBUG nova.scheduler.client.report [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 694.879896] env[65121]: DEBUG nova.compute.provider_tree [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.895156] env[65121]: DEBUG nova.scheduler.client.report [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 694.928075] env[65121]: DEBUG nova.scheduler.client.report [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 695.169726] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.170053] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.205391] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106154, 'name': Rename_Task, 'duration_secs': 0.23534} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.205740] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 695.206079] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a137c98-3776-44ac-bd42-471f8f3d2408 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.217361] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 695.217361] env[65121]: value = "task-5106155" [ 695.217361] env[65121]: _type = "Task" [ 695.217361] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.230953] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.561460] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e39b31f-05f2-45e4-9f97-37a22fba09ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.573725] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e981aa9-5b88-44de-afe5-9f676c0db2cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.619483] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5d0fbe-1f93-4111-b42e-1b9d3664d49a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.630433] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616bd42d-7daf-46d5-a515-9f10a46cb561 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.645748] env[65121]: DEBUG nova.compute.provider_tree [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 695.651709] env[65121]: DEBUG nova.network.neutron [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.730967] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106155, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.956655] env[65121]: DEBUG nova.network.neutron [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Successfully updated port: 0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 696.158365] env[65121]: INFO nova.compute.manager [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Took 2.69 seconds to deallocate network for instance. [ 696.165986] env[65121]: WARNING neutronclient.v2_0.client [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.166999] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.167290] env[65121]: WARNING openstack [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.201917] env[65121]: DEBUG nova.scheduler.client.report [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 51 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 696.202160] env[65121]: DEBUG nova.compute.provider_tree [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 51 to 52 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 696.202378] env[65121]: DEBUG nova.compute.provider_tree [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 696.236390] env[65121]: DEBUG oslo_vmware.api [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106155, 'name': PowerOnVM_Task, 'duration_secs': 0.726446} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.236613] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 696.236975] env[65121]: INFO nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Took 9.54 seconds to spawn the instance on the hypervisor. [ 696.237655] env[65121]: DEBUG nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 696.238701] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6928a9-a4ef-4c0a-8d08-bb04393b3d3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.459777] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.460145] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.460227] env[65121]: DEBUG nova.network.neutron [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 696.679643] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.709100] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.214s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.709300] env[65121]: DEBUG nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 696.714961] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.157s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.716133] env[65121]: INFO nova.compute.claims [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.758523] env[65121]: INFO nova.compute.manager [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Took 31.59 seconds to build instance. [ 696.946801] env[65121]: DEBUG nova.network.neutron [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Updating instance_info_cache with network_info: [{"id": "c2770914-538c-4cb2-9133-0998f766b7ad", "address": "fa:16:3e:3b:da:8f", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2770914-53", "ovs_interfaceid": "c2770914-538c-4cb2-9133-0998f766b7ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.971678] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.971678] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.054739] env[65121]: DEBUG nova.network.neutron [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 697.099045] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.099045] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.206775] env[65121]: WARNING neutronclient.v2_0.client [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.206775] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.206775] env[65121]: WARNING openstack [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.223258] env[65121]: DEBUG nova.compute.utils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 697.227672] env[65121]: DEBUG nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 697.265584] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5dfe6a37-6496-44bb-960a-c36be1a57ce0 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.115s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.372041] env[65121]: DEBUG nova.network.neutron [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 697.452619] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "refresh_cache-27940143-16b5-4263-b23c-354ed8ea8866" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 697.453051] env[65121]: DEBUG nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance network_info: |[{"id": "c2770914-538c-4cb2-9133-0998f766b7ad", "address": "fa:16:3e:3b:da:8f", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2770914-53", "ovs_interfaceid": "c2770914-538c-4cb2-9133-0998f766b7ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 697.453490] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:da:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2770914-538c-4cb2-9133-0998f766b7ad', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 697.461466] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating folder: Project (083c4f2aa7bf4167b472db6bebb9d8bd). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 697.463742] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f8b9f89-e79e-4ab7-b046-a8a65b057f81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.479641] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created folder: Project (083c4f2aa7bf4167b472db6bebb9d8bd) in parent group-v993268. [ 697.479896] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating folder: Instances. Parent ref: group-v993348. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 697.480697] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f1d77cd-007a-499d-a78b-91ac5bb9f121 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.493534] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created folder: Instances in parent group-v993348. [ 697.493775] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 697.493964] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 697.494207] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc20eaad-a42a-4a5c-a7aa-02a4597d09db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.516207] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 697.516207] env[65121]: value = "task-5106158" [ 697.516207] env[65121]: _type = "Task" [ 697.516207] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.526066] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106158, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.583030] env[65121]: DEBUG nova.compute.manager [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Received event network-changed-ea99cbd9-05be-40d9-952c-9bee73993946 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 697.583030] env[65121]: DEBUG nova.compute.manager [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Refreshing instance network info cache due to event network-changed-ea99cbd9-05be-40d9-952c-9bee73993946. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 697.583030] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Acquiring lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.583030] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Acquired lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.583191] env[65121]: DEBUG nova.network.neutron [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Refreshing network info cache for port ea99cbd9-05be-40d9-952c-9bee73993946 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 697.728757] env[65121]: DEBUG nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 697.768569] env[65121]: DEBUG nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 697.876095] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 697.876491] env[65121]: DEBUG nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Instance network_info: |[{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 697.877438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:7e:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e25d678-e865-4027-804f-77e9bb4cb31a', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 697.886030] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating folder: Project (4467da3ed41245ddbc93fc865a8b7bdd). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 697.886742] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0455f2b0-e5c3-4ae8-a155-573be58b0172 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.900173] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Created folder: Project (4467da3ed41245ddbc93fc865a8b7bdd) in parent group-v993268. [ 697.900385] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating folder: Instances. Parent ref: group-v993351. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 697.900698] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af5abb3f-e3aa-44f9-af36-5f882d731dfa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.916023] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Created folder: Instances in parent group-v993351. [ 697.916310] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 697.916508] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 697.917458] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81c28212-4f2b-4b86-90ff-6e57038d2607 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.946219] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 697.946219] env[65121]: value = "task-5106161" [ 697.946219] env[65121]: _type = "Task" [ 697.946219] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.955691] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106161, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.032568] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106158, 'name': CreateVM_Task, 'duration_secs': 0.497129} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.032891] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 698.033886] env[65121]: WARNING neutronclient.v2_0.client [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.034236] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.034850] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.034896] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 698.036596] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d962e49-db50-4052-8a53-8ae366a939d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.046035] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 698.046035] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bb6581-861d-5030-4545-4945635ca3fb" [ 698.046035] env[65121]: _type = "Task" [ 698.046035] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.058618] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bb6581-861d-5030-4545-4945635ca3fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.088216] env[65121]: WARNING neutronclient.v2_0.client [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.088998] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.089495] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.252493] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.253113] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.301842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.368042] env[65121]: WARNING neutronclient.v2_0.client [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.369801] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.370792] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.433289] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09bb2ad-a951-4047-b480-60ad20d0749f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.454776] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170295ad-bba6-4fb1-ad37-3213a4a9ab3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.471156] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106161, 'name': CreateVM_Task, 'duration_secs': 0.47115} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.497277] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 698.499914] env[65121]: DEBUG nova.network.neutron [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Updated VIF entry in instance network info cache for port ea99cbd9-05be-40d9-952c-9bee73993946. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 698.500302] env[65121]: DEBUG nova.network.neutron [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Updating instance_info_cache with network_info: [{"id": "ea99cbd9-05be-40d9-952c-9bee73993946", "address": "fa:16:3e:c1:43:51", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea99cbd9-05", "ovs_interfaceid": "ea99cbd9-05be-40d9-952c-9bee73993946", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 698.503667] env[65121]: WARNING neutronclient.v2_0.client [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.503667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.503667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.503667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 698.503667] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2083481b-cd75-4fe1-8eee-62d5689239af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.507220] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60e37721-a55f-46f5-8478-486ca2f30b34 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.515060] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 698.515060] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ea6ac2-58c7-c34f-663a-8df2b381ccb4" [ 698.515060] env[65121]: _type = "Task" [ 698.515060] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.522193] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b637b613-ea41-44bd-8907-85cd0f467a1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.543332] env[65121]: DEBUG nova.compute.provider_tree [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.550310] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ea6ac2-58c7-c34f-663a-8df2b381ccb4, 'name': SearchDatastore_Task, 'duration_secs': 0.016124} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.558270] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.558270] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.558270] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.558270] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.558270] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.558693] env[65121]: DEBUG nova.compute.manager [req-edd031ca-9a7f-40a1-abbc-95ab366346e9 req-ad8c7aab-fa9b-4d03-94cc-bdeba3bb3a38 service nova] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Received event network-vif-deleted-4f357400-7a42-4e3d-994c-a29afc1b366f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 698.558693] env[65121]: DEBUG nova.compute.manager [req-edd031ca-9a7f-40a1-abbc-95ab366346e9 req-ad8c7aab-fa9b-4d03-94cc-bdeba3bb3a38 service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Received event network-vif-deleted-1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 698.560183] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c0fd2c2-4907-411a-b717-c918acfb4766 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.569853] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bb6581-861d-5030-4545-4945635ca3fb, 'name': SearchDatastore_Task, 'duration_secs': 0.026508} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.570171] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.570405] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.570621] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.570751] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.570918] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.572341] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a62a6a41-86b4-4a69-a23b-1f998543c5c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.574453] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.574642] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 698.575755] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b42c53a8-9ed9-430d-9d66-2508b66a946a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.583575] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 698.583575] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523da6c7-909c-7073-fba3-88a906adceb3" [ 698.583575] env[65121]: _type = "Task" [ 698.583575] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.588484] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.588686] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 698.589814] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5389826a-08ee-4011-9c1c-02f5c2f04dd4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.596495] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523da6c7-909c-7073-fba3-88a906adceb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.601478] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 698.601478] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52626dc6-cb10-ec44-03a8-62d535f76264" [ 698.601478] env[65121]: _type = "Task" [ 698.601478] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.610409] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52626dc6-cb10-ec44-03a8-62d535f76264, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.743632] env[65121]: DEBUG nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 698.775045] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 698.775045] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 698.775045] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 698.775381] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 698.775792] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 698.776049] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 698.776361] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.776719] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 698.779020] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 698.779020] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 698.779020] env[65121]: DEBUG nova.virt.hardware [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 698.779020] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b3e339-3ba1-4b0d-a59c-1f45a4fdb967 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.788151] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ccad981-f29e-4b62-b818-6f0316ccbf7e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.805817] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 698.811960] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Creating folder: Project (0bc48543d199422b9c3224d5602ed961). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 698.812489] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0ba031d-058a-4478-b3c5-8b307e61b6d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.829182] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Created folder: Project (0bc48543d199422b9c3224d5602ed961) in parent group-v993268. [ 698.829182] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Creating folder: Instances. Parent ref: group-v993354. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 698.829182] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f06a23c2-2145-40d8-aad2-2ad8fff0d966 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.840505] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Created folder: Instances in parent group-v993354. [ 698.841745] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 698.841745] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 698.841745] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9887a03-9453-4a43-85c7-15d27a1149d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.864036] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 698.864036] env[65121]: value = "task-5106164" [ 698.864036] env[65121]: _type = "Task" [ 698.864036] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.873887] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106164, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.008663] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Releasing lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.009019] env[65121]: DEBUG nova.compute.manager [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Received event network-changed-1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 699.009613] env[65121]: DEBUG nova.compute.manager [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Refreshing instance network info cache due to event network-changed-1c577738-fcc1-4754-9cee-519ac385ae80. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 699.009613] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Acquiring lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.009613] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Acquired lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.011071] env[65121]: DEBUG nova.network.neutron [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Refreshing network info cache for port 1c577738-fcc1-4754-9cee-519ac385ae80 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 699.053807] env[65121]: DEBUG nova.scheduler.client.report [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 699.102508] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523da6c7-909c-7073-fba3-88a906adceb3, 'name': SearchDatastore_Task, 'duration_secs': 0.013563} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.107015] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03712603-6e7a-4b77-a2ae-15b42a781352 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.117209] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 699.117209] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f00452-d65b-ba8f-d13f-bffba7839a7c" [ 699.117209] env[65121]: _type = "Task" [ 699.117209] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.121161] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52626dc6-cb10-ec44-03a8-62d535f76264, 'name': SearchDatastore_Task, 'duration_secs': 0.023372} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.126536] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5828a335-f584-477c-81e9-fb0d631e824a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.135910] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f00452-d65b-ba8f-d13f-bffba7839a7c, 'name': SearchDatastore_Task, 'duration_secs': 0.01783} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.137735] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.138014] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c/7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.138551] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 699.138551] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bd88cc-b1e5-34bf-5274-49531fae9682" [ 699.138551] env[65121]: _type = "Task" [ 699.138551] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.138729] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a3bbeee-da56-445e-b10d-c9569772059d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.153424] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bd88cc-b1e5-34bf-5274-49531fae9682, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.155230] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 699.155230] env[65121]: value = "task-5106165" [ 699.155230] env[65121]: _type = "Task" [ 699.155230] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.166548] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.375331] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106164, 'name': CreateVM_Task, 'duration_secs': 0.400789} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.375429] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 699.375863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.376022] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.376786] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 699.376786] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ad22858-c118-4605-a719-390ed38ed94f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.383502] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 699.383502] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5283dd90-2cdd-4ff7-c9a6-796edacf6ad5" [ 699.383502] env[65121]: _type = "Task" [ 699.383502] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.393703] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5283dd90-2cdd-4ff7-c9a6-796edacf6ad5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.513872] env[65121]: WARNING neutronclient.v2_0.client [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 699.515406] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.515606] env[65121]: WARNING openstack [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.537853] env[65121]: DEBUG nova.compute.manager [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 699.540050] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3b42b7-9c1e-4c1e-a759-3e9f1d695fcc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.554695] env[65121]: DEBUG nova.network.neutron [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 699.559934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.846s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 699.560725] env[65121]: DEBUG nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 699.565441] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.260s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.568014] env[65121]: INFO nova.compute.claims [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.577156] env[65121]: DEBUG nova.objects.instance [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'flavor' on Instance uuid c4096314-270a-4270-9e1d-5ace8ddbd286 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 699.652825] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bd88cc-b1e5-34bf-5274-49531fae9682, 'name': SearchDatastore_Task, 'duration_secs': 0.015842} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.653290] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.653469] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.653734] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06ff17dd-b59e-4d2b-a9ba-c010da575653 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.666591] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.670845] env[65121]: DEBUG nova.network.neutron [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.672801] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 699.672801] env[65121]: value = "task-5106166" [ 699.672801] env[65121]: _type = "Task" [ 699.672801] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.685470] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106166, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.759688] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.759975] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.895549] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5283dd90-2cdd-4ff7-c9a6-796edacf6ad5, 'name': SearchDatastore_Task, 'duration_secs': 0.014472} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.895851] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.896086] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 699.896310] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.896449] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.896638] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 699.896899] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fd7a215-4579-494b-bbfd-1ef7f5a67e1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.907431] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 699.907607] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 699.908910] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62136289-1fe9-437f-b56f-88f62208fe12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.915262] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 699.915262] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5212317c-58e6-1f33-5872-98d321fad2b1" [ 699.915262] env[65121]: _type = "Task" [ 699.915262] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.923876] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5212317c-58e6-1f33-5872-98d321fad2b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.059622] env[65121]: INFO nova.compute.manager [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] instance snapshotting [ 700.063515] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49eb8ad-b20d-45bc-91b1-c30f69f9f54a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.087979] env[65121]: DEBUG nova.compute.utils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 700.091292] env[65121]: DEBUG nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 700.091492] env[65121]: DEBUG nova.network.neutron [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 700.091812] env[65121]: WARNING neutronclient.v2_0.client [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.092110] env[65121]: WARNING neutronclient.v2_0.client [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.094550] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.094550] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.100627] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b59664-4eb2-42cf-833f-bb707087dbd1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.104045] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.104236] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.104406] env[65121]: DEBUG nova.network.neutron [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 700.104583] env[65121]: DEBUG nova.objects.instance [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'info_cache' on Instance uuid c4096314-270a-4270-9e1d-5ace8ddbd286 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 700.166592] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.170984] env[65121]: DEBUG nova.policy [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff853d2df89c4bc18e74cb1dd62a789f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c0dbd078f6b42d5ad8eeb0a45cac9f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 700.173478] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Releasing lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.173950] env[65121]: DEBUG nova.compute.manager [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Received event network-vif-plugged-c2770914-538c-4cb2-9133-0998f766b7ad {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 700.174245] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Acquiring lock "27940143-16b5-4263-b23c-354ed8ea8866-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.174508] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Lock "27940143-16b5-4263-b23c-354ed8ea8866-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.174724] env[65121]: DEBUG oslo_concurrency.lockutils [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] Lock "27940143-16b5-4263-b23c-354ed8ea8866-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 700.174939] env[65121]: DEBUG nova.compute.manager [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] No waiting events found dispatching network-vif-plugged-c2770914-538c-4cb2-9133-0998f766b7ad {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 700.175177] env[65121]: WARNING nova.compute.manager [req-0ed12d97-b77f-4aa2-a9e3-2453d1ea4d53 req-7a9dc265-cbf0-4bcb-9a07-03af98496a2b service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Received unexpected event network-vif-plugged-c2770914-538c-4cb2-9133-0998f766b7ad for instance with vm_state building and task_state spawning. [ 700.185875] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106166, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.433474] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5212317c-58e6-1f33-5872-98d321fad2b1, 'name': SearchDatastore_Task, 'duration_secs': 0.014732} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.435251] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91ad9683-9f2a-43fe-9d72-6ee2fed92d39 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.447194] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 700.447194] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5269700e-377f-e127-3b7f-cc0f243b8b39" [ 700.447194] env[65121]: _type = "Task" [ 700.447194] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.462890] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5269700e-377f-e127-3b7f-cc0f243b8b39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.521817] env[65121]: DEBUG nova.network.neutron [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Successfully created port: 95f5dff8-7aaa-4dd4-b724-b8f5704e7eff {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 700.592275] env[65121]: DEBUG nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 700.609027] env[65121]: DEBUG nova.objects.base [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 700.619162] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 700.619162] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6c2376da-ba6d-4ab2-9635-0fb932e1268f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.634714] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 700.634714] env[65121]: value = "task-5106167" [ 700.634714] env[65121]: _type = "Task" [ 700.634714] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.647875] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106167, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.675605] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106165, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.691201] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106166, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.960576] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5269700e-377f-e127-3b7f-cc0f243b8b39, 'name': SearchDatastore_Task, 'duration_secs': 0.131095} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.960576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.960576] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 700.960576] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfaf1129-c491-497e-88f8-c6c1cb5bb49d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.970331] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 700.970331] env[65121]: value = "task-5106168" [ 700.970331] env[65121]: _type = "Task" [ 700.970331] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.984566] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106168, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.113211] env[65121]: WARNING neutronclient.v2_0.client [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.113571] env[65121]: WARNING openstack [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.114119] env[65121]: WARNING openstack [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.151122] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106167, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.171623] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106165, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.828821} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.171832] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c/7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 701.172032] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 701.172278] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42bcd551-ef31-42cc-85b0-6baffc0c8b30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.185913] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 701.185913] env[65121]: value = "task-5106169" [ 701.185913] env[65121]: _type = "Task" [ 701.185913] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.194260] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106166, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.282158} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.197851] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 701.198233] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 701.198830] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42822d27-48af-41fb-96e7-fc4297fe0e98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.204803] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106169, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.213678] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 701.213678] env[65121]: value = "task-5106170" [ 701.213678] env[65121]: _type = "Task" [ 701.213678] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.224628] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106170, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.226797] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26697fe3-108d-4ab4-8f10-f15e6b5fb3bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.239451] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5dc3147-fddd-4bfb-b101-377763a17391 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.279708] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff00354f-2bbe-48d0-a8af-d5b12c9f8478 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.292122] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6d8c99-2854-41ed-b2aa-84833fc66174 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.316234] env[65121]: DEBUG nova.compute.provider_tree [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 701.423881] env[65121]: WARNING openstack [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.424513] env[65121]: WARNING openstack [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.487487] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106168, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.606835] env[65121]: DEBUG nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 701.629719] env[65121]: DEBUG nova.compute.manager [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 701.630703] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecaee483-02df-4f8a-a4a6-5aa3316caef4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.646208] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 701.646593] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 701.646669] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 701.646807] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 701.646941] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 701.647090] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 701.647302] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.647460] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 701.647622] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 701.647791] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 701.647954] env[65121]: DEBUG nova.virt.hardware [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 701.649287] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3e5567-e7ba-491a-b0cc-db016ebcc440 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.656082] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106167, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.662457] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7edd174-663f-44f0-8e87-1b7eaf2336b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.699193] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106169, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119001} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.699461] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.700347] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4176134c-7951-4302-b630-f4000c8b9246 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.726977] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c/7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.731618] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dca542fd-6d5c-4105-8f27-7a1af0b51e40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.756080] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106170, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093071} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.758384] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.758974] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 701.758974] env[65121]: value = "task-5106171" [ 701.758974] env[65121]: _type = "Task" [ 701.758974] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.760210] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4693e7b0-72e3-4508-82e0-1fa47adf153e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.813979] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.821176] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e885c27-37e3-4f7c-bd45-2ef1188a1b09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.853552] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106171, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.861134] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 701.861134] env[65121]: value = "task-5106172" [ 701.861134] env[65121]: _type = "Task" [ 701.861134] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.870812] env[65121]: WARNING neutronclient.v2_0.client [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.872095] env[65121]: WARNING openstack [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.872643] env[65121]: WARNING openstack [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.889026] env[65121]: ERROR nova.scheduler.client.report [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [req-3db5ac49-f830-478f-b159-c5d92def8885] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3db5ac49-f830-478f-b159-c5d92def8885"}]} [ 701.898059] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106172, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.914128] env[65121]: DEBUG nova.scheduler.client.report [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 701.935990] env[65121]: DEBUG nova.scheduler.client.report [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 701.936211] env[65121]: DEBUG nova.compute.provider_tree [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 701.951338] env[65121]: DEBUG nova.scheduler.client.report [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 701.983030] env[65121]: DEBUG nova.scheduler.client.report [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 701.990681] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106168, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604684} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.991199] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 701.991600] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 701.991958] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3ca491d-6ffe-4278-b2dd-ea3ef4ced8e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.999972] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 701.999972] env[65121]: value = "task-5106173" [ 701.999972] env[65121]: _type = "Task" [ 701.999972] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.011670] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106173, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.139090] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "293f93f2-c01d-42c8-b1a7-3056805c77de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.139377] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.155677] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106167, 'name': CreateSnapshot_Task, 'duration_secs': 1.353848} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.156727] env[65121]: INFO nova.compute.manager [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] instance snapshotting [ 702.158620] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 702.159847] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b7dfd8-dceb-4029-9e3a-7347904f0f96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.163526] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6067cb1d-803f-44c1-bbe6-e637d7c3bfab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.173769] env[65121]: DEBUG nova.network.neutron [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 702.203315] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.205420] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9f3023-9432-4d79-845d-8e6904b72907 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.211601] env[65121]: DEBUG nova.network.neutron [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Successfully updated port: 95f5dff8-7aaa-4dd4-b724-b8f5704e7eff {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 702.279285] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106171, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.376694] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106172, 'name': ReconfigVM_Task, 'duration_secs': 0.514095} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.376995] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.377776] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ada4b5ce-6b26-4424-8bb1-871bda6187a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.386127] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 702.386127] env[65121]: value = "task-5106174" [ 702.386127] env[65121]: _type = "Task" [ 702.386127] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.396751] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106174, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.511679] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106173, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.165902} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.511947] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 702.512774] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8973a9-824b-42bb-9e78-3934cb482374 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.536135] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 702.539135] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce349b3b-6752-4673-9c1a-6d48a295238f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.561426] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Received event network-changed-c2770914-538c-4cb2-9133-0998f766b7ad {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 702.561426] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Refreshing instance network info cache due to event network-changed-c2770914-538c-4cb2-9133-0998f766b7ad. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 702.561426] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Acquiring lock "refresh_cache-27940143-16b5-4263-b23c-354ed8ea8866" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.561426] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Acquired lock "refresh_cache-27940143-16b5-4263-b23c-354ed8ea8866" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.561426] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Refreshing network info cache for port c2770914-538c-4cb2-9133-0998f766b7ad {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 702.575535] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 702.575535] env[65121]: value = "task-5106175" [ 702.575535] env[65121]: _type = "Task" [ 702.575535] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.587034] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106175, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.644296] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c9dd22-6a3b-ef58-6eb7-989eea130a19/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 702.645413] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19c95f6-9cfc-4ab7-bd69-563cd9fbc4cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.654314] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c9dd22-6a3b-ef58-6eb7-989eea130a19/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 702.654497] env[65121]: ERROR oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c9dd22-6a3b-ef58-6eb7-989eea130a19/disk-0.vmdk due to incomplete transfer. [ 702.654736] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6a7527d6-fb6f-400c-ae8f-be9e604edf33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.657832] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f40f686-6c28-4211-849d-5b7f8d20659c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.667786] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf75232-2ca1-40a0-86f7-928e68507baa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.672099] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c9dd22-6a3b-ef58-6eb7-989eea130a19/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 702.672292] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Uploaded image 45fa7707-1688-4fcb-aa2e-c4460117f6c3 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 702.674502] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 702.675134] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b42e7b72-987e-4483-9786-539d7d6dc19f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.710570] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 702.713988] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5d7074c8-64ea-418c-8a30-bd056ddae369 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.720366] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f55474-9bcf-4f78-9690-61de145986de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.721274] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 702.721274] env[65121]: value = "task-5106176" [ 702.721274] env[65121]: _type = "Task" [ 702.721274] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.722433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "refresh_cache-b1ad9d15-7ef4-404a-9751-ecdd02ea768b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.722433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "refresh_cache-b1ad9d15-7ef4-404a-9751-ecdd02ea768b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.722433] env[65121]: DEBUG nova.network.neutron [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 702.725478] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 702.733619] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-affa0414-b2f0-4a7f-9f1c-9ea144740bca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.739025] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 702.739025] env[65121]: value = "task-5106177" [ 702.739025] env[65121]: _type = "Task" [ 702.739025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.739025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa45c2f-828d-4e1a-b1fe-b79f1a8e4e96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.749038] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106176, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.750725] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 702.750725] env[65121]: value = "task-5106178" [ 702.750725] env[65121]: _type = "Task" [ 702.750725] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.758726] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106177, 'name': CloneVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.767871] env[65121]: DEBUG nova.compute.provider_tree [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 702.775990] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106178, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.783314] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106171, 'name': ReconfigVM_Task, 'duration_secs': 0.553295} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.783314] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Reconfigured VM instance instance-0000001d to attach disk [datastore1] 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c/7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.783565] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-86cdf46d-b7c4-4f6c-bf5d-0bd829b1e510 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.792860] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 702.792860] env[65121]: value = "task-5106179" [ 702.792860] env[65121]: _type = "Task" [ 702.792860] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.808734] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106179, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.900828] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106174, 'name': Rename_Task, 'duration_secs': 0.231945} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.901130] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.901390] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22746bb6-3570-4059-bb32-1542f444889e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.909357] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 702.909357] env[65121]: value = "task-5106180" [ 702.909357] env[65121]: _type = "Task" [ 702.909357] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.920241] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.064481] env[65121]: WARNING neutronclient.v2_0.client [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.064630] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.064966] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.089258] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106175, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.091925] env[65121]: DEBUG nova.compute.manager [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 703.092191] env[65121]: DEBUG nova.compute.manager [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing instance network info cache due to event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 703.092423] env[65121]: DEBUG oslo_concurrency.lockutils [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Acquiring lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.092604] env[65121]: DEBUG oslo_concurrency.lockutils [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Acquired lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.092746] env[65121]: DEBUG nova.network.neutron [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 703.187630] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.188021] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.228287] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 703.228287] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.229211] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.240220] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e007a4cc-0947-4a5b-8848-e853e4ab891e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.253019] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106176, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.259691] env[65121]: DEBUG oslo_vmware.api [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 703.259691] env[65121]: value = "task-5106181" [ 703.259691] env[65121]: _type = "Task" [ 703.259691] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.259884] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106177, 'name': CloneVM_Task} progress is 93%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.266379] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106178, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.274518] env[65121]: DEBUG oslo_vmware.api [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106181, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.282354] env[65121]: WARNING neutronclient.v2_0.client [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.283021] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.283382] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.298192] env[65121]: DEBUG nova.network.neutron [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 703.309197] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106179, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.324560] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.325106] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.333625] env[65121]: DEBUG nova.scheduler.client.report [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 54 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 703.334088] env[65121]: DEBUG nova.compute.provider_tree [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 54 to 55 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 703.334377] env[65121]: DEBUG nova.compute.provider_tree [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.403756] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Updated VIF entry in instance network info cache for port c2770914-538c-4cb2-9133-0998f766b7ad. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 703.404150] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Updating instance_info_cache with network_info: [{"id": "c2770914-538c-4cb2-9133-0998f766b7ad", "address": "fa:16:3e:3b:da:8f", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2770914-53", "ovs_interfaceid": "c2770914-538c-4cb2-9133-0998f766b7ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.407584] env[65121]: WARNING neutronclient.v2_0.client [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.409130] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.409130] env[65121]: WARNING openstack [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.432768] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106180, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.505080] env[65121]: DEBUG nova.network.neutron [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Updating instance_info_cache with network_info: [{"id": "95f5dff8-7aaa-4dd4-b724-b8f5704e7eff", "address": "fa:16:3e:72:9b:8b", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95f5dff8-7a", "ovs_interfaceid": "95f5dff8-7aaa-4dd4-b724-b8f5704e7eff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.590623] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106175, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.597562] env[65121]: WARNING neutronclient.v2_0.client [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.598356] env[65121]: WARNING openstack [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.598757] env[65121]: WARNING openstack [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.721723] env[65121]: WARNING openstack [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.722254] env[65121]: WARNING openstack [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.757231] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106177, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.765012] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106176, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.785724] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106178, 'name': CreateSnapshot_Task, 'duration_secs': 0.622781} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.790677] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 703.791039] env[65121]: DEBUG oslo_vmware.api [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106181, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.791989] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011b8343-9b04-4ee8-a00e-cfc69f484d1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.806820] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106179, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.821590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "19b1f3f9-842e-4150-8890-b0b22393c3af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.821590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.828478] env[65121]: WARNING neutronclient.v2_0.client [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.829255] env[65121]: WARNING openstack [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.829623] env[65121]: WARNING openstack [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.840925] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "61bd5995-701e-430f-9aae-4b266089e313" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.842828] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.278s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.843670] env[65121]: DEBUG nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 703.847771] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.172s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.847771] env[65121]: DEBUG nova.objects.instance [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lazy-loading 'resources' on Instance uuid 6218957b-6329-4004-97ca-07231b55937a {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 703.907627] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Releasing lock "refresh_cache-27940143-16b5-4263-b23c-354ed8ea8866" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.908058] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Received event network-vif-plugged-0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 703.908372] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Acquiring lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.908806] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.910038] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.910126] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] No waiting events found dispatching network-vif-plugged-0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 703.910353] env[65121]: WARNING nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Received unexpected event network-vif-plugged-0e25d678-e865-4027-804f-77e9bb4cb31a for instance with vm_state building and task_state spawning. [ 703.910987] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Received event network-changed-0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 703.910987] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Refreshing instance network info cache due to event network-changed-0e25d678-e865-4027-804f-77e9bb4cb31a. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 703.911288] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Acquiring lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.911640] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Acquired lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.911773] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Refreshing network info cache for port 0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 703.923239] env[65121]: WARNING neutronclient.v2_0.client [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.925075] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.925840] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.948369] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106180, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.966386] env[65121]: DEBUG nova.network.neutron [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updated VIF entry in instance network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 703.966811] env[65121]: DEBUG nova.network.neutron [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updating instance_info_cache with network_info: [{"id": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "address": "fa:16:3e:47:bb:ea", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119a70b-b1", "ovs_interfaceid": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.008059] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "refresh_cache-b1ad9d15-7ef4-404a-9751-ecdd02ea768b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.008763] env[65121]: DEBUG nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Instance network_info: |[{"id": "95f5dff8-7aaa-4dd4-b724-b8f5704e7eff", "address": "fa:16:3e:72:9b:8b", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95f5dff8-7a", "ovs_interfaceid": "95f5dff8-7aaa-4dd4-b724-b8f5704e7eff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 704.009581] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:9b:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95f5dff8-7aaa-4dd4-b724-b8f5704e7eff', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 704.021461] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 704.022231] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 704.022487] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dc43c0c-8cfb-484d-9046-f1090253da63 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.049199] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.049199] env[65121]: value = "task-5106182" [ 704.049199] env[65121]: _type = "Task" [ 704.049199] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.059595] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106182, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.087208] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106175, 'name': ReconfigVM_Task, 'duration_secs': 1.385909} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.087588] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 704.088361] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4674925-a386-4fa7-a807-5e00293ae728 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.093407] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.093800] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.105909] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 704.105909] env[65121]: value = "task-5106183" [ 704.105909] env[65121]: _type = "Task" [ 704.105909] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.120814] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106183, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.192007] env[65121]: WARNING neutronclient.v2_0.client [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.192850] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.193314] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.261194] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106176, 'name': Destroy_Task, 'duration_secs': 1.410127} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.261874] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Destroyed the VM [ 704.262178] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 704.262396] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ee897a46-4ce8-427a-b085-97e47959c17d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.268552] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106177, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.273611] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 704.273611] env[65121]: value = "task-5106184" [ 704.273611] env[65121]: _type = "Task" [ 704.273611] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.281824] env[65121]: DEBUG oslo_vmware.api [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106181, 'name': PowerOnVM_Task, 'duration_secs': 0.660551} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.283805] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 704.283805] env[65121]: DEBUG nova.compute.manager [None req-d256fbd2-0666-49b6-9fcd-150aa3b48c0f tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 704.284105] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd36ebbb-3221-4a66-b0dc-4e26f6b25952 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.290911] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106184, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.301038] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updated VIF entry in instance network info cache for port 0e25d678-e865-4027-804f-77e9bb4cb31a. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 704.301274] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.312413] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106179, 'name': Rename_Task, 'duration_secs': 1.118939} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.312719] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 704.313441] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac33efa3-07a0-4fe6-9b6f-d0657ffb8e3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.328072] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 704.328555] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 704.328555] env[65121]: value = "task-5106185" [ 704.328555] env[65121]: _type = "Task" [ 704.328555] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.329153] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3c66a9b7-438f-4acb-b3bb-e72e7a026025 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.342827] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106185, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.345176] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 704.345176] env[65121]: value = "task-5106186" [ 704.345176] env[65121]: _type = "Task" [ 704.345176] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.351984] env[65121]: DEBUG nova.compute.utils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 704.361314] env[65121]: DEBUG nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 704.361314] env[65121]: DEBUG nova.network.neutron [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 704.361403] env[65121]: WARNING neutronclient.v2_0.client [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.361686] env[65121]: WARNING neutronclient.v2_0.client [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.364722] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.364722] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.370102] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106186, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.417626] env[65121]: DEBUG nova.policy [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc3c22580f4442dcae4a2b5245ede996', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '083c4f2aa7bf4167b472db6bebb9d8bd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 704.434772] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106180, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.469452] env[65121]: DEBUG oslo_concurrency.lockutils [req-67df03b1-35b1-42e9-b718-be5e0a8120c0 req-449b90f4-edc6-4253-99bc-964730ed1578 service nova] Releasing lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.568388] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106182, 'name': CreateVM_Task, 'duration_secs': 0.502582} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.572654] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 704.574711] env[65121]: WARNING neutronclient.v2_0.client [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.574711] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.574711] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.574711] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 704.574996] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08544030-7922-4e68-aa55-273b184cac97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.581660] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 704.581660] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52766736-7490-2bc9-a156-8f0b9069f8a7" [ 704.581660] env[65121]: _type = "Task" [ 704.581660] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.594032] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52766736-7490-2bc9-a156-8f0b9069f8a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.617696] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106183, 'name': Rename_Task, 'duration_secs': 0.206186} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.617935] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 704.618201] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a341bd9-150e-45b3-bd42-bdef919547f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.627048] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 704.627048] env[65121]: value = "task-5106187" [ 704.627048] env[65121]: _type = "Task" [ 704.627048] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.642377] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106187, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.764267] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106177, 'name': CloneVM_Task, 'duration_secs': 1.975206} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.764579] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Created linked-clone VM from snapshot [ 704.765414] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021bf121-7cbf-4624-bcb9-2404dd6f0c51 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.775355] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Uploading image 1a74125e-adbe-4124-bdd0-412251d129e0 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 704.796241] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106184, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.810051] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 704.810051] env[65121]: value = "vm-993358" [ 704.810051] env[65121]: _type = "VirtualMachine" [ 704.810051] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 704.810789] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Releasing lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.811483] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 704.811691] env[65121]: DEBUG nova.compute.manager [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing instance network info cache due to event network-changed-f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 704.811995] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Acquiring lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.812201] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Acquired lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.812398] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Refreshing network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 704.816659] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-54f7dbd4-e679-4347-a7ce-64959e6d2807 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.840277] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lease: (returnval){ [ 704.840277] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52475e0f-4bed-65c1-94cd-07b7c385aa07" [ 704.840277] env[65121]: _type = "HttpNfcLease" [ 704.840277] env[65121]: } obtained for exporting VM: (result){ [ 704.840277] env[65121]: value = "vm-993358" [ 704.840277] env[65121]: _type = "VirtualMachine" [ 704.840277] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 704.840565] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the lease: (returnval){ [ 704.840565] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52475e0f-4bed-65c1-94cd-07b7c385aa07" [ 704.840565] env[65121]: _type = "HttpNfcLease" [ 704.840565] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 704.853676] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106185, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.865550] env[65121]: DEBUG nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 704.875069] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 704.875069] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52475e0f-4bed-65c1-94cd-07b7c385aa07" [ 704.875069] env[65121]: _type = "HttpNfcLease" [ 704.875069] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 704.875382] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106186, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.875655] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 704.875655] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52475e0f-4bed-65c1-94cd-07b7c385aa07" [ 704.875655] env[65121]: _type = "HttpNfcLease" [ 704.875655] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 704.876451] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290964fd-3df0-4f0c-8225-1ba387932f57 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.886748] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523b35ee-bab9-0a31-1668-6d31a5a29da8/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 704.886748] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523b35ee-bab9-0a31-1668-6d31a5a29da8/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 704.955839] env[65121]: DEBUG nova.network.neutron [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Successfully created port: df584c77-b7b3-4d0f-a9eb-6dece1793f87 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 704.965594] env[65121]: DEBUG nova.compute.manager [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Received event network-vif-plugged-95f5dff8-7aaa-4dd4-b724-b8f5704e7eff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 704.965733] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Acquiring lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.965963] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.966165] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.966346] env[65121]: DEBUG nova.compute.manager [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] No waiting events found dispatching network-vif-plugged-95f5dff8-7aaa-4dd4-b724-b8f5704e7eff {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 704.966508] env[65121]: WARNING nova.compute.manager [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Received unexpected event network-vif-plugged-95f5dff8-7aaa-4dd4-b724-b8f5704e7eff for instance with vm_state building and task_state spawning. [ 704.966654] env[65121]: DEBUG nova.compute.manager [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Received event network-changed-95f5dff8-7aaa-4dd4-b724-b8f5704e7eff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 704.966813] env[65121]: DEBUG nova.compute.manager [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Refreshing instance network info cache due to event network-changed-95f5dff8-7aaa-4dd4-b724-b8f5704e7eff. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 704.967010] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Acquiring lock "refresh_cache-b1ad9d15-7ef4-404a-9751-ecdd02ea768b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.967158] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Acquired lock "refresh_cache-b1ad9d15-7ef4-404a-9751-ecdd02ea768b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.967326] env[65121]: DEBUG nova.network.neutron [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Refreshing network info cache for port 95f5dff8-7aaa-4dd4-b724-b8f5704e7eff {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 704.980980] env[65121]: DEBUG oslo_vmware.api [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106180, 'name': PowerOnVM_Task, 'duration_secs': 2.066383} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.985342] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 704.985601] env[65121]: INFO nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Took 13.34 seconds to spawn the instance on the hypervisor. [ 704.986443] env[65121]: DEBUG nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 704.989115] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6882004b-2812-4eea-91e8-05058bc350a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.026048] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-121f3894-7756-4dda-b64f-eb403786644b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.100539] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52766736-7490-2bc9-a156-8f0b9069f8a7, 'name': SearchDatastore_Task, 'duration_secs': 0.018266} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.101473] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.101473] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 705.101473] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.101764] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.101764] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 705.102287] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7205ce4c-73c8-499f-8452-2869ec4ce6df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.116813] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 705.116924] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 705.120510] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1cc8ac2-3307-46f4-aa28-4e37cdd9cf89 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.126612] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 705.126612] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527cd963-9f28-e343-5b36-edefeabcc777" [ 705.126612] env[65121]: _type = "Task" [ 705.126612] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.145837] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106187, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.150700] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527cd963-9f28-e343-5b36-edefeabcc777, 'name': SearchDatastore_Task, 'duration_secs': 0.015701} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.154521] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c18d907a-e7c2-41d3-8a95-861e5edbf6ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.163063] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 705.163063] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529f07ca-f55c-8b82-73c8-ed2fe7d3d24f" [ 705.163063] env[65121]: _type = "Task" [ 705.163063] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.176288] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529f07ca-f55c-8b82-73c8-ed2fe7d3d24f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.178277] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad86bae7-c842-48bf-a093-8977cb87ae2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.186660] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f703126-ac1a-4447-8344-c859eb12708d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.219232] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a7356b-1870-400a-8a30-93bbf72a5bf4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.228435] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f300108-e24f-4161-97c2-57aa55f8829b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.246567] env[65121]: DEBUG nova.compute.provider_tree [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.286425] env[65121]: DEBUG oslo_vmware.api [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106184, 'name': RemoveSnapshot_Task, 'duration_secs': 0.791098} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.286710] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 705.286930] env[65121]: INFO nova.compute.manager [None req-0f5d096f-9fa1-4ade-80ca-ac732e820ba5 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Took 15.95 seconds to snapshot the instance on the hypervisor. [ 705.325095] env[65121]: WARNING neutronclient.v2_0.client [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.325990] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.326430] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.345713] env[65121]: DEBUG oslo_vmware.api [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106185, 'name': PowerOnVM_Task, 'duration_secs': 0.702832} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.346910] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 705.347170] env[65121]: INFO nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Took 10.83 seconds to spawn the instance on the hypervisor. [ 705.347391] env[65121]: DEBUG nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 705.348354] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae94d64-8823-48bc-be03-3870ca15e5e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.369368] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106186, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.470336] env[65121]: WARNING neutronclient.v2_0.client [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.471463] env[65121]: WARNING openstack [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.471948] env[65121]: WARNING openstack [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.508837] env[65121]: INFO nova.compute.manager [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Took 38.43 seconds to build instance. [ 705.646523] env[65121]: DEBUG oslo_vmware.api [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106187, 'name': PowerOnVM_Task, 'duration_secs': 0.577777} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.647048] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 705.647349] env[65121]: INFO nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Took 6.90 seconds to spawn the instance on the hypervisor. [ 705.647718] env[65121]: DEBUG nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 705.648693] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649cf95c-8355-4cc9-ba18-42fb57eea701 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.674670] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529f07ca-f55c-8b82-73c8-ed2fe7d3d24f, 'name': SearchDatastore_Task, 'duration_secs': 0.014611} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.675205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.678023] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b1ad9d15-7ef4-404a-9751-ecdd02ea768b/b1ad9d15-7ef4-404a-9751-ecdd02ea768b.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.678023] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86f7e821-8af1-460c-94e1-d3f3db5e2191 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.685511] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 705.685511] env[65121]: value = "task-5106189" [ 705.685511] env[65121]: _type = "Task" [ 705.685511] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.697063] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106189, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.750498] env[65121]: DEBUG nova.scheduler.client.report [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 705.877268] env[65121]: INFO nova.compute.manager [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Took 37.37 seconds to build instance. [ 705.877765] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106186, 'name': CloneVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.882302] env[65121]: DEBUG nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 705.915248] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 705.915829] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 705.916701] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 705.917043] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 705.917331] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 705.917586] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 705.917957] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.918365] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 705.918674] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 705.919525] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 705.921279] env[65121]: DEBUG nova.virt.hardware [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 705.922434] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8d68f4-de45-4a53-a150-9b5ee4fe7b0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.932477] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f575357d-cb6d-4e28-83e5-5f1678755ce3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.012204] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0908928f-7dba-49ff-9178-9a0db7cbc79c tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "27940143-16b5-4263-b23c-354ed8ea8866" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.941s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.173703] env[65121]: INFO nova.compute.manager [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Took 37.15 seconds to build instance. [ 706.200132] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106189, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.260975] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.414s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.264120] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 28.624s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.264120] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.264120] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 706.264865] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.315s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.265892] env[65121]: INFO nova.compute.claims [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.269259] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67c3453-d5b7-4921-92e1-70aab2ef3a3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.280747] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8fc4c1-1618-47c3-95ed-1b796e4f7382 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.304786] env[65121]: INFO nova.scheduler.client.report [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Deleted allocations for instance 6218957b-6329-4004-97ca-07231b55937a [ 706.307998] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41271711-77ce-4fc2-a7a9-db773d72a377 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.322176] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18773177-7f36-4807-8a84-ba5e6dca516c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.360305] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177634MB free_disk=94GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 706.360452] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.373659] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106186, 'name': CloneVM_Task, 'duration_secs': 1.60511} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.373659] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Created linked-clone VM from snapshot [ 706.374809] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3eb0bd-1cc3-40be-8de5-3fc1c3a59649 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.383840] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2202e0b1-ab35-4dfd-b351-dbaa92b4ecd1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.888s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.385707] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Uploading image f3094449-4692-4fd1-b00e-402a084e36fa {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 706.424337] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 706.424337] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4e644482-8e66-4926-a287-8dcbb8e7e34d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.436308] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 706.436308] env[65121]: value = "task-5106190" [ 706.436308] env[65121]: _type = "Task" [ 706.436308] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.449216] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106190, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.517736] env[65121]: DEBUG nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 706.676314] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c835f2d-a6b8-4078-b908-8b02de1445e3 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.667s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.700626] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106189, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654902} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.701038] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b1ad9d15-7ef4-404a-9751-ecdd02ea768b/b1ad9d15-7ef4-404a-9751-ecdd02ea768b.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.701328] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.701667] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70e27436-8ba3-4fc5-b631-807bbea8b968 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.710832] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 706.710832] env[65121]: value = "task-5106191" [ 706.710832] env[65121]: _type = "Task" [ 706.710832] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.723462] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106191, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.798895] env[65121]: DEBUG nova.network.neutron [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Successfully updated port: df584c77-b7b3-4d0f-a9eb-6dece1793f87 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 706.825673] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ba08b55-be2d-404b-ac25-f63ca2d50488 tempest-InstanceActionsV221TestJSON-1421045681 tempest-InstanceActionsV221TestJSON-1421045681-project-member] Lock "6218957b-6329-4004-97ca-07231b55937a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.706s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.891162] env[65121]: DEBUG nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 706.949444] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106190, 'name': Destroy_Task, 'duration_secs': 0.40495} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.951241] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Destroyed the VM [ 706.951241] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 706.952064] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2d345311-ed52-4994-a30b-b313e0545bcc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.961804] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 706.961804] env[65121]: value = "task-5106192" [ 706.961804] env[65121]: _type = "Task" [ 706.961804] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.972500] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106192, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.060701] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.096841] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.098378] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.116831] env[65121]: WARNING openstack [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.118247] env[65121]: WARNING openstack [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.180207] env[65121]: DEBUG nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 707.237129] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106191, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.303947] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "refresh_cache-9cf6c29a-3424-4b88-9ba5-8120b124beb6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.304401] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "refresh_cache-9cf6c29a-3424-4b88-9ba5-8120b124beb6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.304484] env[65121]: DEBUG nova.network.neutron [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 707.428719] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.470891] env[65121]: WARNING neutronclient.v2_0.client [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.471757] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.472242] env[65121]: WARNING openstack [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.495481] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106192, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.496978] env[65121]: WARNING neutronclient.v2_0.client [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.498360] env[65121]: WARNING openstack [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.498360] env[65121]: WARNING openstack [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.709673] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.730426] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106191, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.74572} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.730833] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 707.732616] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c57beb-aac6-4fca-b132-f5ad2082ea34 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.762069] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] b1ad9d15-7ef4-404a-9751-ecdd02ea768b/b1ad9d15-7ef4-404a-9751-ecdd02ea768b.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.766495] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-313b5548-c7b7-43b9-b6d2-ef8e0382a2de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.791603] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 707.791603] env[65121]: value = "task-5106193" [ 707.791603] env[65121]: _type = "Task" [ 707.791603] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.808174] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.808942] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.821689] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106193, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.837431] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updated VIF entry in instance network info cache for port f119a70b-b122-4875-b1a3-5c4c7a234c29. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 707.837861] env[65121]: DEBUG nova.network.neutron [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updating instance_info_cache with network_info: [{"id": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "address": "fa:16:3e:47:bb:ea", "network": {"id": "d139030f-74bb-44de-934c-f50aeb96b9c4", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-649442954-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e92570886f544d9ea9fe0947e8e93d9c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119a70b-b1", "ovs_interfaceid": "f119a70b-b122-4875-b1a3-5c4c7a234c29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 707.883563] env[65121]: DEBUG nova.network.neutron [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Updated VIF entry in instance network info cache for port 95f5dff8-7aaa-4dd4-b724-b8f5704e7eff. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 707.883951] env[65121]: DEBUG nova.network.neutron [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Updating instance_info_cache with network_info: [{"id": "95f5dff8-7aaa-4dd4-b724-b8f5704e7eff", "address": "fa:16:3e:72:9b:8b", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95f5dff8-7a", "ovs_interfaceid": "95f5dff8-7aaa-4dd4-b724-b8f5704e7eff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 707.991285] env[65121]: DEBUG oslo_vmware.api [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106192, 'name': RemoveSnapshot_Task, 'duration_secs': 0.77807} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.992568] env[65121]: DEBUG nova.network.neutron [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 708.001788] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 708.030051] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401d9a5e-28ff-4712-82a8-762926fd6243 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.041164] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea09d4ef-61b8-4c33-9915-a8ea1421fba7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.047030] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.047374] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.087908] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a007a63-652d-4039-ae73-e3996cac9a94 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.105259] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95f5eef-f69b-45bc-ba84-8b6d2098cae2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.127638] env[65121]: DEBUG nova.compute.provider_tree [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.153876] env[65121]: WARNING neutronclient.v2_0.client [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.154746] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.155157] env[65121]: WARNING openstack [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.303164] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106193, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.341602] env[65121]: DEBUG oslo_concurrency.lockutils [req-6cb82d6a-eeee-498e-ab89-3d70d8e7590f req-c62447e8-c264-4969-a915-4f2992146846 service nova] Releasing lock "refresh_cache-aa9b6708-c53c-4117-9b75-9d506f393395" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.349537] env[65121]: DEBUG nova.network.neutron [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Updating instance_info_cache with network_info: [{"id": "df584c77-b7b3-4d0f-a9eb-6dece1793f87", "address": "fa:16:3e:39:5d:a0", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf584c77-b7", "ovs_interfaceid": "df584c77-b7b3-4d0f-a9eb-6dece1793f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 708.390859] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea5e2d4f-fea5-442b-8102-b20dce2d72bb req-3faffb78-4721-4947-83b8-586d5f9e1e9a service nova] Releasing lock "refresh_cache-b1ad9d15-7ef4-404a-9751-ecdd02ea768b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.507372] env[65121]: WARNING nova.compute.manager [None req-e6f5dff6-11fd-435a-ba55-e9753ace1cf7 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Image not found during snapshot: nova.exception.ImageNotFound: Image f3094449-4692-4fd1-b00e-402a084e36fa could not be found. [ 708.630783] env[65121]: DEBUG nova.scheduler.client.report [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 708.714389] env[65121]: DEBUG nova.compute.manager [req-4dd4c742-efc8-4e95-97ad-39ef6f346e10 req-0fe7c9df-7212-4f4b-a34c-8b184841634f service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Received event network-vif-plugged-df584c77-b7b3-4d0f-a9eb-6dece1793f87 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 708.714389] env[65121]: DEBUG oslo_concurrency.lockutils [req-4dd4c742-efc8-4e95-97ad-39ef6f346e10 req-0fe7c9df-7212-4f4b-a34c-8b184841634f service nova] Acquiring lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.714890] env[65121]: DEBUG oslo_concurrency.lockutils [req-4dd4c742-efc8-4e95-97ad-39ef6f346e10 req-0fe7c9df-7212-4f4b-a34c-8b184841634f service nova] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.714890] env[65121]: DEBUG oslo_concurrency.lockutils [req-4dd4c742-efc8-4e95-97ad-39ef6f346e10 req-0fe7c9df-7212-4f4b-a34c-8b184841634f service nova] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 708.715045] env[65121]: DEBUG nova.compute.manager [req-4dd4c742-efc8-4e95-97ad-39ef6f346e10 req-0fe7c9df-7212-4f4b-a34c-8b184841634f service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] No waiting events found dispatching network-vif-plugged-df584c77-b7b3-4d0f-a9eb-6dece1793f87 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 708.715110] env[65121]: WARNING nova.compute.manager [req-4dd4c742-efc8-4e95-97ad-39ef6f346e10 req-0fe7c9df-7212-4f4b-a34c-8b184841634f service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Received unexpected event network-vif-plugged-df584c77-b7b3-4d0f-a9eb-6dece1793f87 for instance with vm_state building and task_state spawning. [ 708.802294] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106193, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.853608] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "refresh_cache-9cf6c29a-3424-4b88-9ba5-8120b124beb6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.853921] env[65121]: DEBUG nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Instance network_info: |[{"id": "df584c77-b7b3-4d0f-a9eb-6dece1793f87", "address": "fa:16:3e:39:5d:a0", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf584c77-b7", "ovs_interfaceid": "df584c77-b7b3-4d0f-a9eb-6dece1793f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 708.854499] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:5d:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df584c77-b7b3-4d0f-a9eb-6dece1793f87', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.867190] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 708.867190] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 708.867576] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b65fb69-3e99-4234-a3ac-12d1638d0eee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.899123] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.899123] env[65121]: value = "task-5106194" [ 708.899123] env[65121]: _type = "Task" [ 708.899123] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.909621] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106194, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.137013] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.873s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.137590] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 709.140411] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.643s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.141840] env[65121]: INFO nova.compute.claims [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.305435] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106193, 'name': ReconfigVM_Task, 'duration_secs': 1.140147} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.305774] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Reconfigured VM instance instance-0000001f to attach disk [datastore1] b1ad9d15-7ef4-404a-9751-ecdd02ea768b/b1ad9d15-7ef4-404a-9751-ecdd02ea768b.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 709.306531] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9558655c-a8c3-45b5-9fbc-7c69d462fe29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.308949] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "6a4841b6-6076-4e9c-a6cf-a658957e2931" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.309250] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.309537] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "6a4841b6-6076-4e9c-a6cf-a658957e2931-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.309750] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.310699] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.313044] env[65121]: INFO nova.compute.manager [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Terminating instance [ 709.318703] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 709.318703] env[65121]: value = "task-5106195" [ 709.318703] env[65121]: _type = "Task" [ 709.318703] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.337992] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106195, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.410914] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106194, 'name': CreateVM_Task, 'duration_secs': 0.4697} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.411284] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 709.411604] env[65121]: WARNING neutronclient.v2_0.client [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 709.411969] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.412178] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.412507] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 709.412767] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da3eb97a-f5a3-4a2a-a12b-681dd32e8796 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.419116] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 709.419116] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527bb802-feeb-039e-a763-50140ac7b9a0" [ 709.419116] env[65121]: _type = "Task" [ 709.419116] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.429298] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527bb802-feeb-039e-a763-50140ac7b9a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.648438] env[65121]: DEBUG nova.compute.utils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 709.653868] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 709.653868] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 709.653868] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 709.653868] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 709.654124] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.654501] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 709.715457] env[65121]: DEBUG nova.policy [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea485ec50b264dfeb9e93ff0ae5a7544', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f0b5fd1c88e44d3952279db674447a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 709.821234] env[65121]: DEBUG nova.compute.manager [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 709.821234] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.822078] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916c481e-5d8b-4dc6-93bb-28d524202913 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.835183] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106195, 'name': Rename_Task, 'duration_secs': 0.221135} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.837821] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 709.838153] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 709.838419] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d141f4a1-1037-4c10-86c8-d425ff82ab94 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.840473] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65184d24-189a-4e59-8a2f-eb818550b9b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.849024] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 709.849024] env[65121]: value = "task-5106196" [ 709.849024] env[65121]: _type = "Task" [ 709.849024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.861551] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.934965] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527bb802-feeb-039e-a763-50140ac7b9a0, 'name': SearchDatastore_Task, 'duration_secs': 0.013146} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.934965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.934965] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.934965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.934965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.934965] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.936843] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1b4ccec-6d61-4bbc-b538-1fd3a04dedf9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.939925] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 709.940258] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 709.940537] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Deleting the datastore file [datastore2] 6a4841b6-6076-4e9c-a6cf-a658957e2931 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 709.941051] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d7e0925-f19f-4ce8-a4d9-98a01cd14655 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.952316] env[65121]: DEBUG oslo_vmware.api [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 709.952316] env[65121]: value = "task-5106198" [ 709.952316] env[65121]: _type = "Task" [ 709.952316] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.956076] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.956369] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 709.957554] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b87b123-a3a0-4f08-8eb0-c0387fb2366b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.965024] env[65121]: DEBUG oslo_vmware.api [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106198, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.969027] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 709.969027] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e2b40d-be1a-aa77-794c-ec38d3e9eb57" [ 709.969027] env[65121]: _type = "Task" [ 709.969027] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.980447] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e2b40d-be1a-aa77-794c-ec38d3e9eb57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.154368] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 710.201137] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Successfully created port: 97194a81-8ff7-43dd-83f6-c4cce03f1985 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 710.372866] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106196, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.462904] env[65121]: DEBUG oslo_vmware.api [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106198, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319953} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.463291] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 710.465129] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 710.465129] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 710.465129] env[65121]: INFO nova.compute.manager [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Took 0.64 seconds to destroy the instance on the hypervisor. [ 710.465129] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 710.465129] env[65121]: DEBUG nova.compute.manager [-] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 710.465129] env[65121]: DEBUG nova.network.neutron [-] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 710.465129] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 710.465129] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.465666] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.486454] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e2b40d-be1a-aa77-794c-ec38d3e9eb57, 'name': SearchDatastore_Task, 'duration_secs': 0.017131} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.487327] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07cbf19c-9d6d-4b7a-82a5-d7624c87f131 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.494562] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 710.494562] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526f5879-fca5-aa49-d3ab-2bb55fe8edbd" [ 710.494562] env[65121]: _type = "Task" [ 710.494562] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.513503] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526f5879-fca5-aa49-d3ab-2bb55fe8edbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.851069] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0dab3c1-62c9-4779-b887-4fe6abbae946 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.867827] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08d84c9-dc11-4230-937f-b0e07a099150 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.871650] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106196, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.909426] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8a3166-6351-4d12-baac-c58ea2506b44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.918483] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99beea02-9d8e-476e-8750-4f608f3a2abb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.934215] env[65121]: DEBUG nova.compute.provider_tree [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.011205] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526f5879-fca5-aa49-d3ab-2bb55fe8edbd, 'name': SearchDatastore_Task, 'duration_secs': 0.016915} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.011467] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.011820] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 9cf6c29a-3424-4b88-9ba5-8120b124beb6/9cf6c29a-3424-4b88-9ba5-8120b124beb6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 711.012257] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94ea7dac-a5fb-478b-918d-6babb779776e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.022199] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 711.022199] env[65121]: value = "task-5106199" [ 711.022199] env[65121]: _type = "Task" [ 711.022199] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.032600] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.171616] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 711.203492] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 711.203728] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 711.203871] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 711.204053] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 711.204626] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 711.204626] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 711.204737] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.204923] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 711.205105] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 711.205272] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 711.205467] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 711.206874] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5ebba4-17ea-4a70-adcf-0232e251c35c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.217793] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da8a32a-ea0b-4cf6-9af4-8706533f037e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.370741] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106196, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.437906] env[65121]: DEBUG nova.scheduler.client.report [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 711.535658] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106199, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.872857] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106196, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.872857] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Successfully updated port: 97194a81-8ff7-43dd-83f6-c4cce03f1985 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 711.946245] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.805s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.949793] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 711.951304] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.297s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.951587] env[65121]: DEBUG nova.objects.instance [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lazy-loading 'resources' on Instance uuid 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 711.963986] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.970312] env[65121]: DEBUG nova.objects.instance [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lazy-loading 'flavor' on Instance uuid 5854d21a-d1a8-4043-aec8-b37ff25c40e7 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 712.039179] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724795} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.039441] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 9cf6c29a-3424-4b88-9ba5-8120b124beb6/9cf6c29a-3424-4b88-9ba5-8120b124beb6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 712.039658] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.039936] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-407dc2c5-0e5c-42f3-8881-706a4e4ee45c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.048215] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 712.048215] env[65121]: value = "task-5106200" [ 712.048215] env[65121]: _type = "Task" [ 712.048215] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.057689] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.366656] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106196, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.378494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "refresh_cache-6164d64c-75c5-469e-ac32-667842b3717a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.378494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "refresh_cache-6164d64c-75c5-469e-ac32-667842b3717a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 712.378494] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 712.460162] env[65121]: DEBUG nova.compute.utils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 712.461286] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 712.461484] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 712.461912] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.462104] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.462688] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.463044] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.482824] env[65121]: DEBUG oslo_concurrency.lockutils [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.483027] env[65121]: DEBUG oslo_concurrency.lockutils [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquired lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 712.486579] env[65121]: WARNING neutronclient.v2_0.client [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.487441] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.487808] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.522763] env[65121]: DEBUG nova.policy [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea485ec50b264dfeb9e93ff0ae5a7544', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f0b5fd1c88e44d3952279db674447a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 712.566278] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077685} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.566606] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 712.567405] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f3d29c-319a-4a09-9cf6-19c2f357875a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.597384] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Reconfiguring VM instance instance-00000020 to attach disk [datastore1] 9cf6c29a-3424-4b88-9ba5-8120b124beb6/9cf6c29a-3424-4b88-9ba5-8120b124beb6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 712.600203] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ce3f247-8725-45d3-9117-15ecc07007b0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.622388] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 712.622388] env[65121]: value = "task-5106201" [ 712.622388] env[65121]: _type = "Task" [ 712.622388] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.633837] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106201, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.762432] env[65121]: DEBUG nova.network.neutron [-] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 712.867202] env[65121]: DEBUG oslo_vmware.api [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106196, 'name': PowerOnVM_Task, 'duration_secs': 2.545613} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.870687] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 712.870800] env[65121]: INFO nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Took 11.26 seconds to spawn the instance on the hypervisor. [ 712.870979] env[65121]: DEBUG nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 712.872544] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffd12c5-b413-4007-bbb1-2058c2215a40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.882600] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Successfully created port: 38e0b295-fa6e-4faa-be7b-8ec82a785df2 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 712.885542] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.885887] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.933744] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 712.974502] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 713.136793] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106201, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.162494] env[65121]: DEBUG nova.network.neutron [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 713.192610] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.193271] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.209382] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73d4bc7-a86a-4453-bac2-1bca1fd20995 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.221349] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6a3caf-b21f-499c-bc23-051588cfe5d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.262060] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af69517-16c9-4406-a31e-dd310913a364 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.266019] env[65121]: INFO nova.compute.manager [-] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Took 2.80 seconds to deallocate network for instance. [ 713.276274] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e6d7ab-c1ef-4162-a3bf-bf2794110f7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.301143] env[65121]: DEBUG nova.compute.provider_tree [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.411406] env[65121]: INFO nova.compute.manager [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Took 41.89 seconds to build instance. [ 713.474403] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.475176] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.475550] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.635162] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106201, 'name': ReconfigVM_Task, 'duration_secs': 0.645531} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.635162] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Reconfigured VM instance instance-00000020 to attach disk [datastore1] 9cf6c29a-3424-4b88-9ba5-8120b124beb6/9cf6c29a-3424-4b88-9ba5-8120b124beb6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 713.635162] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c083aa0e-43be-4130-bc77-ad58880e901b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.642633] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 713.642633] env[65121]: value = "task-5106202" [ 713.642633] env[65121]: _type = "Task" [ 713.642633] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.653310] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106202, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.672947] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Updating instance_info_cache with network_info: [{"id": "97194a81-8ff7-43dd-83f6-c4cce03f1985", "address": "fa:16:3e:53:e2:f2", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97194a81-8f", "ovs_interfaceid": "97194a81-8ff7-43dd-83f6-c4cce03f1985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 713.675481] env[65121]: WARNING neutronclient.v2_0.client [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.675817] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.676199] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.761107] env[65121]: DEBUG nova.compute.manager [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Received event network-changed-df584c77-b7b3-4d0f-a9eb-6dece1793f87 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 713.761107] env[65121]: DEBUG nova.compute.manager [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Refreshing instance network info cache due to event network-changed-df584c77-b7b3-4d0f-a9eb-6dece1793f87. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 713.761107] env[65121]: DEBUG oslo_concurrency.lockutils [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Acquiring lock "refresh_cache-9cf6c29a-3424-4b88-9ba5-8120b124beb6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.761107] env[65121]: DEBUG oslo_concurrency.lockutils [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Acquired lock "refresh_cache-9cf6c29a-3424-4b88-9ba5-8120b124beb6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.761107] env[65121]: DEBUG nova.network.neutron [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Refreshing network info cache for port df584c77-b7b3-4d0f-a9eb-6dece1793f87 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 713.779591] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 713.786140] env[65121]: DEBUG nova.compute.manager [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Received event network-changed-0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 713.786140] env[65121]: DEBUG nova.compute.manager [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Refreshing instance network info cache due to event network-changed-0e25d678-e865-4027-804f-77e9bb4cb31a. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 713.786140] env[65121]: DEBUG oslo_concurrency.lockutils [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Acquiring lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.786140] env[65121]: DEBUG oslo_concurrency.lockutils [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Acquired lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.786140] env[65121]: DEBUG nova.network.neutron [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Refreshing network info cache for port 0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 713.805733] env[65121]: DEBUG nova.scheduler.client.report [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 713.904792] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.905434] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.916163] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8e7c2cf8-12a2-4d1b-b95b-be90551494f7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.120s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.991767] env[65121]: WARNING neutronclient.v2_0.client [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.993608] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.994239] env[65121]: WARNING openstack [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.008881] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 714.041332] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 714.041613] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 714.041784] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 714.042799] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 714.042799] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 714.042799] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 714.042799] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.042988] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 714.043629] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 714.043780] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 714.043976] env[65121]: DEBUG nova.virt.hardware [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 714.045488] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72989eb1-5db5-4100-b257-e7bf30df4a46 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.059654] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b7b099-5666-4c10-8eba-0a4c1c232ba7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.153052] env[65121]: DEBUG nova.network.neutron [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 714.159135] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106202, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.185809] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "refresh_cache-6164d64c-75c5-469e-ac32-667842b3717a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.186191] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Instance network_info: |[{"id": "97194a81-8ff7-43dd-83f6-c4cce03f1985", "address": "fa:16:3e:53:e2:f2", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97194a81-8f", "ovs_interfaceid": "97194a81-8ff7-43dd-83f6-c4cce03f1985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 714.186647] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:e2:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97194a81-8ff7-43dd-83f6-c4cce03f1985', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.196476] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 714.196476] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 714.196476] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6ef2528-e1fd-4061-9ca5-18744d65221c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.219593] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.219593] env[65121]: value = "task-5106203" [ 714.219593] env[65121]: _type = "Task" [ 714.219593] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.232623] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106203, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.266122] env[65121]: WARNING neutronclient.v2_0.client [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.267065] env[65121]: WARNING openstack [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.267480] env[65121]: WARNING openstack [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.289054] env[65121]: WARNING neutronclient.v2_0.client [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.290098] env[65121]: WARNING openstack [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.290864] env[65121]: WARNING openstack [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.310826] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.359s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.316731] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.515s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.316731] env[65121]: INFO nova.compute.claims [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.340425] env[65121]: INFO nova.scheduler.client.report [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted allocations for instance 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4 [ 714.403504] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.403692] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.404670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.404670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.404670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.406879] env[65121]: INFO nova.compute.manager [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Terminating instance [ 714.419662] env[65121]: DEBUG nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 714.500418] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "aab9317b-4ee6-48b3-905b-859a5996f33d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.500665] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.608095] env[65121]: WARNING openstack [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.608622] env[65121]: WARNING openstack [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.650626] env[65121]: WARNING openstack [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.652123] env[65121]: WARNING openstack [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.660321] env[65121]: DEBUG oslo_concurrency.lockutils [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Releasing lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.660576] env[65121]: DEBUG nova.compute.manager [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Inject network info {{(pid=65121) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 714.660963] env[65121]: DEBUG nova.compute.manager [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] network_info to inject: |[{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 714.668565] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Reconfiguring VM instance to set the machine id {{(pid=65121) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 714.675056] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Successfully updated port: 38e0b295-fa6e-4faa-be7b-8ec82a785df2 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 714.677387] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec06eb07-c07c-447c-b76f-6134dab65a26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.694687] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106202, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.695713] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "refresh_cache-ab27fa56-f672-4096-a8f4-cce5ff4d5460" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.695827] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "refresh_cache-ab27fa56-f672-4096-a8f4-cce5ff4d5460" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.696772] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 714.704719] env[65121]: DEBUG oslo_vmware.api [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 714.704719] env[65121]: value = "task-5106204" [ 714.704719] env[65121]: _type = "Task" [ 714.704719] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.716769] env[65121]: DEBUG oslo_vmware.api [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106204, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.735017] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106203, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.851806] env[65121]: DEBUG oslo_concurrency.lockutils [None req-316d7a2f-1a01-440a-b665-784af6675426 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "2700ee2c-cc57-40e0-8069-f1b7ad2df7b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.936s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.862204] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "46f89cfb-14aa-483d-aa8a-bc91356272ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.862500] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.862685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "46f89cfb-14aa-483d-aa8a-bc91356272ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.862862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.863025] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.867088] env[65121]: INFO nova.compute.manager [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Terminating instance [ 714.913082] env[65121]: DEBUG nova.compute.manager [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 714.913082] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 714.914719] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230a2d84-2afe-46d4-84c1-0a503244b053 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.922825] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 714.923171] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0d91a2a-fd00-43eb-b935-375325d22f37 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.943113] env[65121]: DEBUG oslo_vmware.api [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 714.943113] env[65121]: value = "task-5106205" [ 714.943113] env[65121]: _type = "Task" [ 714.943113] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.955425] env[65121]: WARNING neutronclient.v2_0.client [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.955425] env[65121]: WARNING openstack [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.955546] env[65121]: WARNING openstack [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.970042] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.970517] env[65121]: DEBUG oslo_vmware.api [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106205, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.990927] env[65121]: WARNING neutronclient.v2_0.client [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.991856] env[65121]: WARNING openstack [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.992525] env[65121]: WARNING openstack [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.160195] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106202, 'name': Rename_Task, 'duration_secs': 1.311845} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.168421] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 715.168421] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5eef2f7a-137c-4148-a43b-1cc07b9f20e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.180046] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 715.180046] env[65121]: value = "task-5106206" [ 715.180046] env[65121]: _type = "Task" [ 715.180046] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.196999] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.199036] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.200581] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.227763] env[65121]: DEBUG oslo_vmware.api [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106204, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.242547] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106203, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.264034] env[65121]: DEBUG nova.network.neutron [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Updated VIF entry in instance network info cache for port df584c77-b7b3-4d0f-a9eb-6dece1793f87. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 715.264034] env[65121]: DEBUG nova.network.neutron [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Updating instance_info_cache with network_info: [{"id": "df584c77-b7b3-4d0f-a9eb-6dece1793f87", "address": "fa:16:3e:39:5d:a0", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf584c77-b7", "ovs_interfaceid": "df584c77-b7b3-4d0f-a9eb-6dece1793f87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 715.327378] env[65121]: DEBUG nova.network.neutron [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updated VIF entry in instance network info cache for port 0e25d678-e865-4027-804f-77e9bb4cb31a. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 715.327843] env[65121]: DEBUG nova.network.neutron [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 715.357370] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "75114f97-fe50-4624-9333-303e411529ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 715.357370] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "75114f97-fe50-4624-9333-303e411529ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 715.374987] env[65121]: DEBUG nova.compute.manager [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 715.375366] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.377174] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672827a4-e394-4d14-9353-c3a6aa4536ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.382796] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 715.395714] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 715.396252] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b357ce2-ce95-4a59-ae47-569fdecdaf0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.406136] env[65121]: DEBUG oslo_vmware.api [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 715.406136] env[65121]: value = "task-5106207" [ 715.406136] env[65121]: _type = "Task" [ 715.406136] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.419959] env[65121]: DEBUG oslo_vmware.api [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106207, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.432170] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.432170] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.453944] env[65121]: DEBUG oslo_vmware.api [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106205, 'name': PowerOffVM_Task, 'duration_secs': 0.304122} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.454219] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 715.454401] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 715.454662] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50e05b2c-5380-4809-b367-8599f09be65e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.509808] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 715.509808] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.510025] env[65121]: WARNING openstack [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.539098] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 715.539344] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 715.539573] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleting the datastore file [datastore2] 7a0d0769-31c5-4ae1-8520-a744f64d39c6 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 715.539794] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7d59bf7-a433-4e39-ba70-6839ffb2a465 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.554788] env[65121]: DEBUG oslo_vmware.api [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 715.554788] env[65121]: value = "task-5106209" [ 715.554788] env[65121]: _type = "Task" [ 715.554788] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.565321] env[65121]: DEBUG oslo_vmware.api [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106209, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.693786] env[65121]: DEBUG nova.network.neutron [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Updating instance_info_cache with network_info: [{"id": "38e0b295-fa6e-4faa-be7b-8ec82a785df2", "address": "fa:16:3e:ae:d4:7a", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e0b295-fa", "ovs_interfaceid": "38e0b295-fa6e-4faa-be7b-8ec82a785df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 715.701803] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106206, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.725219] env[65121]: DEBUG oslo_vmware.api [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106204, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.738954] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106203, 'name': CreateVM_Task, 'duration_secs': 1.048625} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.739150] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 715.742711] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 715.743153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.743315] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 715.743759] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 715.744176] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea8b7301-217c-4259-ae27-c9de02910f76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.751445] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 715.751445] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5244d80a-2b17-f9e3-243a-cd1cd12fa435" [ 715.751445] env[65121]: _type = "Task" [ 715.751445] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.765038] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5244d80a-2b17-f9e3-243a-cd1cd12fa435, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.765624] env[65121]: DEBUG oslo_concurrency.lockutils [req-44df4f28-45f2-4c9e-8b45-2cfea1aa181e req-e45a3d57-5af3-4a4e-9c05-1ae430bae52c service nova] Releasing lock "refresh_cache-9cf6c29a-3424-4b88-9ba5-8120b124beb6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.838162] env[65121]: DEBUG oslo_concurrency.lockutils [req-7eed5f8c-476a-475c-aee9-df45d3f816ca req-40b592c8-d8c3-4aa5-a400-1c69b54653c9 service nova] Releasing lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.922879] env[65121]: DEBUG oslo_vmware.api [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106207, 'name': PowerOffVM_Task, 'duration_secs': 0.303352} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.922879] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 715.922879] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 715.922879] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1ff8c38-e8e8-4740-b190-83213abf8c71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.006364] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 716.006582] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 716.006731] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Deleting the datastore file [datastore2] 46f89cfb-14aa-483d-aa8a-bc91356272ca {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 716.007313] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f92f0b8-2208-4ce5-9c9d-196f0398f75f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.017195] env[65121]: DEBUG oslo_vmware.api [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for the task: (returnval){ [ 716.017195] env[65121]: value = "task-5106211" [ 716.017195] env[65121]: _type = "Task" [ 716.017195] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.025297] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523b35ee-bab9-0a31-1668-6d31a5a29da8/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 716.026321] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e5ab48-09e8-45d5-ad49-431bbc21cf4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.033640] env[65121]: DEBUG oslo_vmware.api [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106211, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.035161] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874bc346-6481-458e-9be1-866a7e8bfdf9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.043113] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523b35ee-bab9-0a31-1668-6d31a5a29da8/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 716.043458] env[65121]: ERROR oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523b35ee-bab9-0a31-1668-6d31a5a29da8/disk-0.vmdk due to incomplete transfer. [ 716.046017] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3d76b8c6-ccd0-4307-878c-5a1ac31a7a4a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.048676] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27011d01-0080-4d54-b9ef-3c7682827d2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.092045] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf4d84f-d684-4950-97e7-110966fb3f38 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.095317] env[65121]: DEBUG oslo_vmware.rw_handles [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523b35ee-bab9-0a31-1668-6d31a5a29da8/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 716.095587] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Uploaded image 1a74125e-adbe-4124-bdd0-412251d129e0 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 716.097914] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 716.101140] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c995001e-7d9c-4f73-a670-7fba86dc31fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.102917] env[65121]: DEBUG oslo_vmware.api [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106209, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283822} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.103922] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 716.106024] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 716.106024] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.106024] env[65121]: INFO nova.compute.manager [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Took 1.19 seconds to destroy the instance on the hypervisor. [ 716.106024] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 716.106024] env[65121]: DEBUG nova.compute.manager [-] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 716.106024] env[65121]: DEBUG nova.network.neutron [-] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 716.106024] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.106573] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.106839] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.122795] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 716.122795] env[65121]: value = "task-5106212" [ 716.122795] env[65121]: _type = "Task" [ 716.122795] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.125163] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22c502f-2e16-4c1f-810a-285e34cf68a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.152469] env[65121]: DEBUG nova.compute.provider_tree [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 716.159476] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106212, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.166604] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.196023] env[65121]: DEBUG oslo_vmware.api [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106206, 'name': PowerOnVM_Task, 'duration_secs': 0.79226} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.196023] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 716.196023] env[65121]: INFO nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Took 10.31 seconds to spawn the instance on the hypervisor. [ 716.196023] env[65121]: DEBUG nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 716.196023] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6e4798-95a6-44e9-9ed8-b3293d465893 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.197269] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "refresh_cache-ab27fa56-f672-4096-a8f4-cce5ff4d5460" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.197722] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Instance network_info: |[{"id": "38e0b295-fa6e-4faa-be7b-8ec82a785df2", "address": "fa:16:3e:ae:d4:7a", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e0b295-fa", "ovs_interfaceid": "38e0b295-fa6e-4faa-be7b-8ec82a785df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 716.198166] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:d4:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38e0b295-fa6e-4faa-be7b-8ec82a785df2', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.205730] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 716.206748] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 716.207019] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aaf064ff-4692-4f8c-8501-e35c0087ba48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.236210] env[65121]: DEBUG oslo_vmware.api [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106204, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.237719] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.237719] env[65121]: value = "task-5106213" [ 716.237719] env[65121]: _type = "Task" [ 716.237719] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.247347] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106213, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.263868] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5244d80a-2b17-f9e3-243a-cd1cd12fa435, 'name': SearchDatastore_Task, 'duration_secs': 0.023669} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.263868] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.263868] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 716.264145] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.264145] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.264851] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.264851] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8ac39e9-332d-45aa-9f27-25a697a99d99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.281165] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.281165] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 716.281165] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4e2a496-de29-4641-904f-c28030b2e9e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.291045] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 716.291045] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527f898f-de27-e46b-c6cf-75376df663ae" [ 716.291045] env[65121]: _type = "Task" [ 716.291045] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.301173] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527f898f-de27-e46b-c6cf-75376df663ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.531215] env[65121]: DEBUG oslo_vmware.api [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Task: {'id': task-5106211, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361654} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.531493] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 716.531674] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 716.531841] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.532013] env[65121]: INFO nova.compute.manager [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Took 1.16 seconds to destroy the instance on the hypervisor. [ 716.532372] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 716.532596] env[65121]: DEBUG nova.compute.manager [-] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 716.532691] env[65121]: DEBUG nova.network.neutron [-] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 716.532950] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.533512] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.533770] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.602842] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.643629] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106212, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.676622] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Received event network-vif-plugged-97194a81-8ff7-43dd-83f6-c4cce03f1985 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 716.677487] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Acquiring lock "6164d64c-75c5-469e-ac32-667842b3717a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.677487] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Lock "6164d64c-75c5-469e-ac32-667842b3717a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.677487] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Lock "6164d64c-75c5-469e-ac32-667842b3717a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 716.677823] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] No waiting events found dispatching network-vif-plugged-97194a81-8ff7-43dd-83f6-c4cce03f1985 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 716.677823] env[65121]: WARNING nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Received unexpected event network-vif-plugged-97194a81-8ff7-43dd-83f6-c4cce03f1985 for instance with vm_state building and task_state spawning. [ 716.678140] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Received event network-changed-97194a81-8ff7-43dd-83f6-c4cce03f1985 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 716.678228] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Refreshing instance network info cache due to event network-changed-97194a81-8ff7-43dd-83f6-c4cce03f1985. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 716.678482] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Acquiring lock "refresh_cache-6164d64c-75c5-469e-ac32-667842b3717a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.678596] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Acquired lock "refresh_cache-6164d64c-75c5-469e-ac32-667842b3717a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.678846] env[65121]: DEBUG nova.network.neutron [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Refreshing network info cache for port 97194a81-8ff7-43dd-83f6-c4cce03f1985 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 716.694576] env[65121]: ERROR nova.scheduler.client.report [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [req-883bb039-0e0c-495a-9ce0-8f24a3c5cdec] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-883bb039-0e0c-495a-9ce0-8f24a3c5cdec"}]} [ 716.719388] env[65121]: DEBUG nova.scheduler.client.report [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 716.750241] env[65121]: DEBUG nova.scheduler.client.report [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 716.750446] env[65121]: DEBUG nova.compute.provider_tree [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 716.756685] env[65121]: DEBUG oslo_vmware.api [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106204, 'name': ReconfigVM_Task, 'duration_secs': 1.605157} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.757944] env[65121]: INFO nova.compute.manager [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Took 40.51 seconds to build instance. [ 716.763733] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-36ed852b-2da0-4eab-a39c-5e6f422d684e tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Reconfigured VM instance to set the machine id {{(pid=65121) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 716.769523] env[65121]: DEBUG nova.compute.manager [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Received event network-changed-5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 716.770056] env[65121]: DEBUG nova.compute.manager [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing instance network info cache due to event network-changed-5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 716.770056] env[65121]: DEBUG oslo_concurrency.lockutils [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Acquiring lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.773765] env[65121]: DEBUG oslo_concurrency.lockutils [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Acquired lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.773765] env[65121]: DEBUG nova.network.neutron [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 716.780359] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106213, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.789123] env[65121]: DEBUG nova.scheduler.client.report [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 716.806589] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527f898f-de27-e46b-c6cf-75376df663ae, 'name': SearchDatastore_Task, 'duration_secs': 0.026672} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.808652] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3ce0854-37fb-4659-9bad-32f0002c5525 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.823634] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 716.823634] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528aba4b-f4e4-7616-4ba2-33d6e6b27f1f" [ 716.823634] env[65121]: _type = "Task" [ 716.823634] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.836902] env[65121]: DEBUG nova.scheduler.client.report [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 716.848009] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528aba4b-f4e4-7616-4ba2-33d6e6b27f1f, 'name': SearchDatastore_Task, 'duration_secs': 0.019016} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.849415] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.849415] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 6164d64c-75c5-469e-ac32-667842b3717a/6164d64c-75c5-469e-ac32-667842b3717a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 716.849415] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb1fe397-4d3f-44ad-891a-3147008d58f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.857589] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 716.857589] env[65121]: value = "task-5106214" [ 716.857589] env[65121]: _type = "Task" [ 716.857589] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.870897] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106214, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.994011] env[65121]: DEBUG nova.network.neutron [-] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 717.145648] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106212, 'name': Destroy_Task, 'duration_secs': 0.725925} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.151293] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Destroyed the VM [ 717.151892] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 717.154090] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8bee85e9-be9f-48e6-8857-e8cd582c2a4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.164700] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 717.164700] env[65121]: value = "task-5106215" [ 717.164700] env[65121]: _type = "Task" [ 717.164700] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.188961] env[65121]: WARNING neutronclient.v2_0.client [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.190378] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.190911] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.208903] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106215, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.250919] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106213, 'name': CreateVM_Task, 'duration_secs': 0.602754} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.251147] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 717.256089] env[65121]: WARNING neutronclient.v2_0.client [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.256519] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.256811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.257054] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 717.257662] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-671031f8-0bee-4b7b-a7fe-99db221cc798 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.265055] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 717.265055] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52daee9a-823b-cc9f-c92b-daff077874c6" [ 717.265055] env[65121]: _type = "Task" [ 717.265055] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.272734] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1fac7d5-0acc-4cab-957d-47f46cbb9539 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.445s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.293599] env[65121]: WARNING neutronclient.v2_0.client [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.295041] env[65121]: WARNING openstack [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.295392] env[65121]: WARNING openstack [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.306173] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52daee9a-823b-cc9f-c92b-daff077874c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.344224] env[65121]: INFO nova.compute.manager [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Rebuilding instance [ 717.379688] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106214, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.407974] env[65121]: DEBUG nova.compute.manager [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 717.408994] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717fc450-9e9c-4cf3-9fa0-2d08b91de499 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.469261] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.469640] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.477306] env[65121]: DEBUG nova.network.neutron [-] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 717.496882] env[65121]: INFO nova.compute.manager [-] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Took 1.39 seconds to deallocate network for instance. [ 717.679908] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106215, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.764751] env[65121]: WARNING neutronclient.v2_0.client [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.765451] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.765815] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.779130] env[65121]: DEBUG nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 717.788961] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52daee9a-823b-cc9f-c92b-daff077874c6, 'name': SearchDatastore_Task, 'duration_secs': 0.076026} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.792054] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 717.794267] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.794267] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.794267] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.794267] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.794267] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de647b94-29af-4e18-b4ce-71903c1befab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.805909] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.805909] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 717.807056] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb562f8b-d4ad-464a-b123-9092119d9364 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.813128] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 717.813128] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a0f206-1dfb-cd2e-7371-b15af4b19fc4" [ 717.813128] env[65121]: _type = "Task" [ 717.813128] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.822608] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a0f206-1dfb-cd2e-7371-b15af4b19fc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.829429] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62b3680-3a92-4675-8170-5d926fa1e161 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.840366] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14275e44-dd45-4698-ba01-389acb80d1a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.846017] env[65121]: WARNING openstack [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.846017] env[65121]: WARNING openstack [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.894027] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf61e04-0a6c-47f8-bee7-4523f379a16b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.909857] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb05f09-e442-44ab-abb3-8519c10bab47 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.913033] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106214, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695188} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.913148] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 6164d64c-75c5-469e-ac32-667842b3717a/6164d64c-75c5-469e-ac32-667842b3717a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 717.913976] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.914577] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb642dfa-bde3-4241-a4dc-9652c6f62cb8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.925820] env[65121]: DEBUG nova.compute.provider_tree [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.942343] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 717.942343] env[65121]: value = "task-5106216" [ 717.942343] env[65121]: _type = "Task" [ 717.942343] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.954834] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106216, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.982417] env[65121]: INFO nova.compute.manager [-] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Took 1.45 seconds to deallocate network for instance. [ 718.006428] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.180971] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106215, 'name': RemoveSnapshot_Task} progress is 31%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.256616] env[65121]: DEBUG nova.network.neutron [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Updated VIF entry in instance network info cache for port 97194a81-8ff7-43dd-83f6-c4cce03f1985. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 718.257039] env[65121]: DEBUG nova.network.neutron [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Updating instance_info_cache with network_info: [{"id": "97194a81-8ff7-43dd-83f6-c4cce03f1985", "address": "fa:16:3e:53:e2:f2", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97194a81-8f", "ovs_interfaceid": "97194a81-8ff7-43dd-83f6-c4cce03f1985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 718.299386] env[65121]: WARNING neutronclient.v2_0.client [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.302213] env[65121]: WARNING openstack [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.302213] env[65121]: WARNING openstack [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.320905] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.328265] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a0f206-1dfb-cd2e-7371-b15af4b19fc4, 'name': SearchDatastore_Task, 'duration_secs': 0.022881} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.329132] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58fe2940-a481-42da-9b58-13388d87e5c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.337162] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 718.337162] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5215fcd9-d74b-97c1-5b13-1b2ec547ce5e" [ 718.337162] env[65121]: _type = "Task" [ 718.337162] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.348373] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5215fcd9-d74b-97c1-5b13-1b2ec547ce5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.429533] env[65121]: DEBUG nova.scheduler.client.report [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 718.437208] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 718.437208] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-073e0a48-1eae-43d0-95da-2374aec4de8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.451974] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 718.451974] env[65121]: value = "task-5106217" [ 718.451974] env[65121]: _type = "Task" [ 718.451974] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.469714] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106216, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076035} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.474109] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.474742] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.475710] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ecef2b3-c4b8-4b3a-ba47-3cf310333f0f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.498188] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.508302] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 6164d64c-75c5-469e-ac32-667842b3717a/6164d64c-75c5-469e-ac32-667842b3717a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.508659] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c6d5b3b-7996-4613-a530-0963918676bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.531918] env[65121]: DEBUG nova.network.neutron [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updated VIF entry in instance network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 718.532405] env[65121]: DEBUG nova.network.neutron [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 718.539366] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 718.539366] env[65121]: value = "task-5106218" [ 718.539366] env[65121]: _type = "Task" [ 718.539366] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.557421] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106218, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.680150] env[65121]: DEBUG oslo_vmware.api [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106215, 'name': RemoveSnapshot_Task, 'duration_secs': 1.509254} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.680428] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 718.680724] env[65121]: INFO nova.compute.manager [None req-63978389-3b30-4ab5-95a6-5419968552ae tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Took 18.62 seconds to snapshot the instance on the hypervisor. [ 718.761264] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Releasing lock "refresh_cache-6164d64c-75c5-469e-ac32-667842b3717a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.761594] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Received event network-vif-deleted-18509dce-9df7-41cc-af3e-70f419fbb5d5 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 718.761848] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Received event network-vif-plugged-38e0b295-fa6e-4faa-be7b-8ec82a785df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 718.762083] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Acquiring lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.762357] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.762659] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.762867] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] No waiting events found dispatching network-vif-plugged-38e0b295-fa6e-4faa-be7b-8ec82a785df2 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 718.763163] env[65121]: WARNING nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Received unexpected event network-vif-plugged-38e0b295-fa6e-4faa-be7b-8ec82a785df2 for instance with vm_state building and task_state spawning. [ 718.763749] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Received event network-changed-38e0b295-fa6e-4faa-be7b-8ec82a785df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 718.763749] env[65121]: DEBUG nova.compute.manager [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Refreshing instance network info cache due to event network-changed-38e0b295-fa6e-4faa-be7b-8ec82a785df2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 718.763749] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Acquiring lock "refresh_cache-ab27fa56-f672-4096-a8f4-cce5ff4d5460" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.763880] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Acquired lock "refresh_cache-ab27fa56-f672-4096-a8f4-cce5ff4d5460" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.764159] env[65121]: DEBUG nova.network.neutron [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Refreshing network info cache for port 38e0b295-fa6e-4faa-be7b-8ec82a785df2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 718.852959] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5215fcd9-d74b-97c1-5b13-1b2ec547ce5e, 'name': SearchDatastore_Task, 'duration_secs': 0.05994} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.853402] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.853402] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ab27fa56-f672-4096-a8f4-cce5ff4d5460/ab27fa56-f672-4096-a8f4-cce5ff4d5460.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 718.854869] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de9327a1-3656-4758-b7fb-454d7dd23644 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.862168] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 718.862168] env[65121]: value = "task-5106219" [ 718.862168] env[65121]: _type = "Task" [ 718.862168] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.873984] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.936033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.622s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.936729] env[65121]: DEBUG nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 718.939815] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.873s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.942303] env[65121]: DEBUG nova.objects.instance [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lazy-loading 'resources' on Instance uuid b9c78b3b-dd43-4199-98a7-ff57fd129f43 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 718.972373] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106217, 'name': PowerOffVM_Task, 'duration_secs': 0.452719} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.972373] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 718.972373] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.972970] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f687daeb-8193-40ed-8a67-e6e24a5de952 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.982243] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 718.982578] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09638b97-2731-4a00-a5fb-02f815999a7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.010352] env[65121]: DEBUG nova.objects.instance [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lazy-loading 'flavor' on Instance uuid 5854d21a-d1a8-4043-aec8-b37ff25c40e7 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 719.036568] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 719.036568] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 719.036568] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Deleting the datastore file [datastore1] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 719.036921] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8486029-9a27-4509-bdd6-4b812e9180ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.040144] env[65121]: DEBUG oslo_concurrency.lockutils [req-35a3b688-936b-4557-a761-db546219bd2e req-3f8df440-4fd7-43af-ac8a-1cd3b664182d service nova] Releasing lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.059925] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106218, 'name': ReconfigVM_Task, 'duration_secs': 0.523564} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.062283] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 6164d64c-75c5-469e-ac32-667842b3717a/6164d64c-75c5-469e-ac32-667842b3717a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.064020] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 719.064020] env[65121]: value = "task-5106221" [ 719.064020] env[65121]: _type = "Task" [ 719.064020] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.064020] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-479b5913-8c70-49c2-bbee-a5e7d0aa6454 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.078050] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106221, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.078795] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 719.078795] env[65121]: value = "task-5106222" [ 719.078795] env[65121]: _type = "Task" [ 719.078795] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.090273] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106222, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.267609] env[65121]: WARNING neutronclient.v2_0.client [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.268363] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.268685] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.378357] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106219, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.389785] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.390341] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.438518] env[65121]: DEBUG nova.compute.manager [req-d527ce5e-9c7e-4807-ab95-0fbdb1cc8f35 req-b312318d-1d4b-4c4c-86cd-2887c5846b92 service nova] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Received event network-vif-deleted-33c6c3ae-feb6-4512-ac82-46c27afad318 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 719.438518] env[65121]: DEBUG nova.compute.manager [req-d527ce5e-9c7e-4807-ab95-0fbdb1cc8f35 req-b312318d-1d4b-4c4c-86cd-2887c5846b92 service nova] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Received event network-vif-deleted-16314fe2-22b5-437b-aebf-58bf042902d0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 719.448885] env[65121]: DEBUG nova.compute.utils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 719.456530] env[65121]: DEBUG nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 719.457480] env[65121]: DEBUG nova.network.neutron [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 719.458117] env[65121]: WARNING neutronclient.v2_0.client [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.458372] env[65121]: WARNING neutronclient.v2_0.client [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.458989] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.459370] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.471106] env[65121]: WARNING neutronclient.v2_0.client [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.471763] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.472165] env[65121]: WARNING openstack [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.517111] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.517662] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquired lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.517662] env[65121]: WARNING neutronclient.v2_0.client [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.532124] env[65121]: DEBUG nova.policy [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff853d2df89c4bc18e74cb1dd62a789f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c0dbd078f6b42d5ad8eeb0a45cac9f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 719.582336] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106221, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176395} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.585523] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 719.585715] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 719.585885] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.595063] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106222, 'name': Rename_Task, 'duration_secs': 0.476243} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.597872] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 719.600017] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5d4f04b-5815-4047-a035-32a738d6aefc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.606413] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 719.606413] env[65121]: value = "task-5106223" [ 719.606413] env[65121]: _type = "Task" [ 719.606413] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.618141] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106223, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.878992] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106219, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647198} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.879519] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ab27fa56-f672-4096-a8f4-cce5ff4d5460/ab27fa56-f672-4096-a8f4-cce5ff4d5460.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 719.879730] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 719.880035] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce7a9966-a493-4fd3-b2f8-781ad0e475b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.886815] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 719.886815] env[65121]: value = "task-5106224" [ 719.886815] env[65121]: _type = "Task" [ 719.886815] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.899650] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106224, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.957747] env[65121]: DEBUG nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 720.121568] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106223, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.122965] env[65121]: DEBUG nova.network.neutron [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Successfully created port: 2cd8c718-cf4c-4d35-92ae-d95f45242770 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 720.188828] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c89b649-7a35-4410-9312-09fcb648b9de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.198882] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76bc719-510a-4051-a9cb-2d80bb6a145a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.245950] env[65121]: DEBUG nova.network.neutron [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Updated VIF entry in instance network info cache for port 38e0b295-fa6e-4faa-be7b-8ec82a785df2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 720.246256] env[65121]: DEBUG nova.network.neutron [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Updating instance_info_cache with network_info: [{"id": "38e0b295-fa6e-4faa-be7b-8ec82a785df2", "address": "fa:16:3e:ae:d4:7a", "network": {"id": "900c858d-e629-4f7f-b301-68de5133ff7e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1424851941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f0b5fd1c88e44d3952279db674447a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e0b295-fa", "ovs_interfaceid": "38e0b295-fa6e-4faa-be7b-8ec82a785df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 720.248036] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60571aa3-c588-4215-b7df-5eba77ddd3f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.258352] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51223d4-3540-4539-908f-38633cbb8d10 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.275954] env[65121]: DEBUG nova.compute.provider_tree [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.399436] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106224, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072725} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.399548] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 720.400267] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0f8e55-38bc-47e0-8bea-e5eba2e52b65 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.423993] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] ab27fa56-f672-4096-a8f4-cce5ff4d5460/ab27fa56-f672-4096-a8f4-cce5ff4d5460.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 720.426043] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb901ca0-6b6f-46d2-90c1-769e311c6c67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.449508] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 720.449508] env[65121]: value = "task-5106225" [ 720.449508] env[65121]: _type = "Task" [ 720.449508] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.461477] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106225, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.619932] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106223, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.632954] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 720.633254] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 720.633413] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 720.633602] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 720.633791] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 720.633957] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 720.634219] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.634434] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 720.634620] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 720.634786] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 720.634979] env[65121]: DEBUG nova.virt.hardware [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 720.636021] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be28ae09-7871-46be-95ed-e585137404bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.647700] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e00df3e-002a-4eba-9668-4dee2d121560 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.665321] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.671858] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 720.672185] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 720.672493] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46494716-43fc-4b06-ac36-573d719c0a3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.694024] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.694024] env[65121]: value = "task-5106226" [ 720.694024] env[65121]: _type = "Task" [ 720.694024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.703936] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106226, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.752424] env[65121]: DEBUG oslo_concurrency.lockutils [req-195f1878-1c86-4400-b41a-101d502f7d97 req-b9966a08-c047-4282-a110-3a9dbe05d63d service nova] Releasing lock "refresh_cache-ab27fa56-f672-4096-a8f4-cce5ff4d5460" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.769087] env[65121]: DEBUG nova.network.neutron [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 720.780618] env[65121]: DEBUG nova.scheduler.client.report [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 720.964432] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106225, 'name': ReconfigVM_Task, 'duration_secs': 0.494857} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.964856] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Reconfigured VM instance instance-00000022 to attach disk [datastore2] ab27fa56-f672-4096-a8f4-cce5ff4d5460/ab27fa56-f672-4096-a8f4-cce5ff4d5460.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.965718] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98459b8c-01e6-4633-b75d-d9e8f0aa4df7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.969337] env[65121]: DEBUG nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 720.980924] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 720.980924] env[65121]: value = "task-5106227" [ 720.980924] env[65121]: _type = "Task" [ 720.980924] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.992117] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106227, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.005554] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 721.005781] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 721.005924] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 721.006281] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 721.006528] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 721.006943] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 721.007585] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.007873] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 721.008491] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 721.008712] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 721.009083] env[65121]: DEBUG nova.virt.hardware [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 721.012617] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8687d1-e71b-4d0c-8c2f-0ca0ed380b21 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.025734] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f030ac90-97a5-41e9-933d-09770eeff3a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.120018] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106223, 'name': PowerOnVM_Task, 'duration_secs': 1.088224} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.120294] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 721.120489] env[65121]: INFO nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Took 9.95 seconds to spawn the instance on the hypervisor. [ 721.120663] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 721.121474] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2844f18d-0e3b-4336-a02c-ef0c50999fbf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.215305] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106226, 'name': CreateVM_Task, 'duration_secs': 0.312925} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.215305] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 721.215669] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.216192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.216192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 721.216506] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d261d1e4-b963-499d-b8f6-9309f075374d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.223505] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 721.223505] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525e062b-ff05-318e-83c6-49b09ec18b23" [ 721.223505] env[65121]: _type = "Task" [ 721.223505] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.234704] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525e062b-ff05-318e-83c6-49b09ec18b23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.275709] env[65121]: WARNING neutronclient.v2_0.client [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.276737] env[65121]: WARNING openstack [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.278113] env[65121]: WARNING openstack [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.293586] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.354s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.296334] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.181s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.298125] env[65121]: INFO nova.compute.claims [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.348701] env[65121]: INFO nova.scheduler.client.report [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Deleted allocations for instance b9c78b3b-dd43-4199-98a7-ff57fd129f43 [ 721.444499] env[65121]: WARNING openstack [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.444912] env[65121]: WARNING openstack [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.492506] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106227, 'name': Rename_Task, 'duration_secs': 0.2279} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.493026] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 721.493475] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5c5dc80-fd3e-46f2-839a-53120b3ad155 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.505228] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 721.505228] env[65121]: value = "task-5106228" [ 721.505228] env[65121]: _type = "Task" [ 721.505228] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.518239] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.641791] env[65121]: INFO nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Took 43.72 seconds to build instance. [ 721.737794] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525e062b-ff05-318e-83c6-49b09ec18b23, 'name': SearchDatastore_Task, 'duration_secs': 0.023122} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.737951] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.738430] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 721.738537] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.739067] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.739067] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 721.739238] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3dc36628-104b-40cf-9c9d-76ab47197206 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.751046] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 721.751046] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 721.751243] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1dbaa94-670d-493a-8adf-74e30f3d593d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.757372] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 721.757372] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52809446-db09-936d-6931-42eaf76994f2" [ 721.757372] env[65121]: _type = "Task" [ 721.757372] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.774108] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52809446-db09-936d-6931-42eaf76994f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.867943] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd1bbbb-b841-48c0-af61-0a41aaa5fd96 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "b9c78b3b-dd43-4199-98a7-ff57fd129f43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.165s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.958087] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "b6f6871d-311c-4adb-824e-2907a12f4224" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.958087] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "b6f6871d-311c-4adb-824e-2907a12f4224" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 722.019259] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106228, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.046557] env[65121]: WARNING neutronclient.v2_0.client [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.047402] env[65121]: WARNING openstack [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.047827] env[65121]: WARNING openstack [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.145044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "6164d64c-75c5-469e-ac32-667842b3717a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.926s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 722.184688] env[65121]: DEBUG nova.network.neutron [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Successfully updated port: 2cd8c718-cf4c-4d35-92ae-d95f45242770 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 722.269596] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52809446-db09-936d-6931-42eaf76994f2, 'name': SearchDatastore_Task, 'duration_secs': 0.013025} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.275797] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48fe6aa6-25e0-462f-a907-e004dd8b7c68 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.282370] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 722.282370] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52922518-8102-9886-88f0-e554fe273938" [ 722.282370] env[65121]: _type = "Task" [ 722.282370] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.292555] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52922518-8102-9886-88f0-e554fe273938, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.357781] env[65121]: DEBUG nova.network.neutron [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 722.521165] env[65121]: DEBUG oslo_vmware.api [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106228, 'name': PowerOnVM_Task, 'duration_secs': 0.774474} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.523998] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 722.524362] env[65121]: INFO nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Took 8.52 seconds to spawn the instance on the hypervisor. [ 722.524542] env[65121]: DEBUG nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 722.525749] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139dfb19-5628-476e-970f-189f971d1bef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.646822] env[65121]: DEBUG nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 722.700667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "refresh_cache-c1004635-b318-489c-9e16-6cb545279953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.700667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "refresh_cache-c1004635-b318-489c-9e16-6cb545279953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.700667] env[65121]: DEBUG nova.network.neutron [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 722.796931] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52922518-8102-9886-88f0-e554fe273938, 'name': SearchDatastore_Task, 'duration_secs': 0.011384} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.800736] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.801155] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 722.802405] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18ceaecf-a3e1-45a2-b958-72beb8e97422 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.811437] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 722.811437] env[65121]: value = "task-5106229" [ 722.811437] env[65121]: _type = "Task" [ 722.811437] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.827715] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106229, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.861211] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Releasing lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.862156] env[65121]: DEBUG nova.compute.manager [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Inject network info {{(pid=65121) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 722.862156] env[65121]: DEBUG nova.compute.manager [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] network_info to inject: |[{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 722.866768] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Reconfiguring VM instance to set the machine id {{(pid=65121) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 722.867134] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-036d1b09-a3be-41a3-82eb-20077d3a25bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.883802] env[65121]: DEBUG oslo_vmware.api [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 722.883802] env[65121]: value = "task-5106230" [ 722.883802] env[65121]: _type = "Task" [ 722.883802] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.893882] env[65121]: DEBUG oslo_vmware.api [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106230, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.976804] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de655571-5f2f-4923-8481-58244724c727 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.985393] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ade48e7-fe13-4cc7-ab07-f24e6c5c55c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.019191] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83ec623-d7ef-49ec-b574-f19baaf260e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.028698] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbb6560-1376-43de-ae5f-b39fa8aa2a29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.058926] env[65121]: DEBUG nova.compute.provider_tree [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.061569] env[65121]: INFO nova.compute.manager [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Took 43.59 seconds to build instance. [ 723.179739] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.203563] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.204278] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.273318] env[65121]: DEBUG nova.network.neutron [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 723.301642] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.301642] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.333706] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106229, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.403293] env[65121]: DEBUG oslo_vmware.api [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106230, 'name': ReconfigVM_Task, 'duration_secs': 0.216615} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.403584] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1834fc72-71e2-408d-b832-85b11fc59657 tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Reconfigured VM instance to set the machine id {{(pid=65121) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 723.449877] env[65121]: DEBUG nova.compute.manager [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Received event network-changed-5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 723.450176] env[65121]: DEBUG nova.compute.manager [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing instance network info cache due to event network-changed-5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 723.450176] env[65121]: DEBUG oslo_concurrency.lockutils [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Acquiring lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.450439] env[65121]: DEBUG oslo_concurrency.lockutils [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Acquired lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.450439] env[65121]: DEBUG nova.network.neutron [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Refreshing network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 723.459791] env[65121]: WARNING neutronclient.v2_0.client [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.460128] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.460492] env[65121]: WARNING openstack [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.563431] env[65121]: DEBUG nova.scheduler.client.report [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 723.569885] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c17aa2b-84fa-4ee9-a9ce-9bd68f0ece92 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.320s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.596785] env[65121]: DEBUG oslo_concurrency.lockutils [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Acquiring lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.596863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.608436] env[65121]: DEBUG nova.network.neutron [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Updating instance_info_cache with network_info: [{"id": "2cd8c718-cf4c-4d35-92ae-d95f45242770", "address": "fa:16:3e:ff:2e:63", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd8c718-cf", "ovs_interfaceid": "2cd8c718-cf4c-4d35-92ae-d95f45242770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 723.619475] env[65121]: DEBUG nova.compute.manager [req-d7e849be-dc58-4147-8060-bdfcd5495ed3 req-96b6cacd-6dd2-4011-9626-7715c195eddc service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Received event network-vif-plugged-2cd8c718-cf4c-4d35-92ae-d95f45242770 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 723.619475] env[65121]: DEBUG oslo_concurrency.lockutils [req-d7e849be-dc58-4147-8060-bdfcd5495ed3 req-96b6cacd-6dd2-4011-9626-7715c195eddc service nova] Acquiring lock "c1004635-b318-489c-9e16-6cb545279953-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.619475] env[65121]: DEBUG oslo_concurrency.lockutils [req-d7e849be-dc58-4147-8060-bdfcd5495ed3 req-96b6cacd-6dd2-4011-9626-7715c195eddc service nova] Lock "c1004635-b318-489c-9e16-6cb545279953-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.619475] env[65121]: DEBUG oslo_concurrency.lockutils [req-d7e849be-dc58-4147-8060-bdfcd5495ed3 req-96b6cacd-6dd2-4011-9626-7715c195eddc service nova] Lock "c1004635-b318-489c-9e16-6cb545279953-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.619475] env[65121]: DEBUG nova.compute.manager [req-d7e849be-dc58-4147-8060-bdfcd5495ed3 req-96b6cacd-6dd2-4011-9626-7715c195eddc service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] No waiting events found dispatching network-vif-plugged-2cd8c718-cf4c-4d35-92ae-d95f45242770 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 723.619475] env[65121]: WARNING nova.compute.manager [req-d7e849be-dc58-4147-8060-bdfcd5495ed3 req-96b6cacd-6dd2-4011-9626-7715c195eddc service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Received unexpected event network-vif-plugged-2cd8c718-cf4c-4d35-92ae-d95f45242770 for instance with vm_state building and task_state spawning. [ 723.824242] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106229, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681569} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.825104] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 723.825104] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.825225] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22a9aba9-6305-420c-bf63-ad992477d0b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.836019] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 723.836019] env[65121]: value = "task-5106231" [ 723.836019] env[65121]: _type = "Task" [ 723.836019] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.848854] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106231, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.932998] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "00362477-c89e-4f60-98a4-d4928081d55e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.933271] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "00362477-c89e-4f60-98a4-d4928081d55e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.954124] env[65121]: WARNING neutronclient.v2_0.client [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.958023] env[65121]: WARNING openstack [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.958023] env[65121]: WARNING openstack [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.074768] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.778s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.075573] env[65121]: DEBUG nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 724.081133] env[65121]: DEBUG nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 724.085350] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.150s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 724.085350] env[65121]: DEBUG nova.objects.instance [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lazy-loading 'resources' on Instance uuid 124876fe-a4f2-4140-8dee-d8bb72b5c9a5 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 724.100826] env[65121]: DEBUG nova.compute.utils [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 724.112308] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "refresh_cache-c1004635-b318-489c-9e16-6cb545279953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.112586] env[65121]: DEBUG nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Instance network_info: |[{"id": "2cd8c718-cf4c-4d35-92ae-d95f45242770", "address": "fa:16:3e:ff:2e:63", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd8c718-cf", "ovs_interfaceid": "2cd8c718-cf4c-4d35-92ae-d95f45242770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 724.113550] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:2e:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cd8c718-cf4c-4d35-92ae-d95f45242770', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.127290] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 724.127290] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1004635-b318-489c-9e16-6cb545279953] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 724.130321] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b14f30e8-3f14-44af-b45f-84c843d43cab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.158055] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.158055] env[65121]: value = "task-5106232" [ 724.158055] env[65121]: _type = "Task" [ 724.158055] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.174761] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106232, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.207403] env[65121]: WARNING openstack [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.207797] env[65121]: WARNING openstack [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.290447] env[65121]: WARNING neutronclient.v2_0.client [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.291339] env[65121]: WARNING openstack [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.291908] env[65121]: WARNING openstack [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.318827] env[65121]: DEBUG nova.compute.manager [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 724.319817] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f059c59-ba63-4dc3-9401-50176919ed56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.346959] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106231, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091906} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.346959] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 724.347712] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953ee14e-6987-4ff6-9472-f7f7d6dacdbf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.372720] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 724.377684] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abc7d8db-c0ff-4e4f-87bc-aeb82b5336c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.401794] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 724.401794] env[65121]: value = "task-5106233" [ 724.401794] env[65121]: _type = "Task" [ 724.401794] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.417044] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106233, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.440706] env[65121]: DEBUG nova.network.neutron [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updated VIF entry in instance network info cache for port 5604925f-4cea-418a-a42d-502bac3a6114. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 724.441057] env[65121]: DEBUG nova.network.neutron [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [{"id": "5604925f-4cea-418a-a42d-502bac3a6114", "address": "fa:16:3e:e0:3e:a1", "network": {"id": "1bdbf860-cd4a-42e7-8aac-dfe2f7092c32", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-196952852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ede261894564e9d9bf50ce9cddeec77", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9abd00f-2cea-40f8-9804-a56b6431192d", "external-id": "nsx-vlan-transportzone-639", "segmentation_id": 639, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5604925f-4c", "ovs_interfaceid": "5604925f-4cea-418a-a42d-502bac3a6114", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.586272] env[65121]: DEBUG nova.compute.utils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 724.587591] env[65121]: DEBUG nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 724.587774] env[65121]: DEBUG nova.network.neutron [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 724.588106] env[65121]: WARNING neutronclient.v2_0.client [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.588399] env[65121]: WARNING neutronclient.v2_0.client [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.589052] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.589390] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.604562] env[65121]: DEBUG oslo_concurrency.lockutils [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.624855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 724.655997] env[65121]: DEBUG nova.policy [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '53e291faf6804d8693617896d260a6d2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f0e2921e8704997ba4a1f5cf982989f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 724.674482] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106232, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.832089] env[65121]: INFO nova.compute.manager [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] instance snapshotting [ 724.835434] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00e3ee3-4908-4564-844f-5f0133ee2fa4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.864664] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cf932b-b6f4-467b-9209-8c22e33e1bef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.922914] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.945292] env[65121]: DEBUG oslo_concurrency.lockutils [req-6ba3d5b0-5eb2-4c48-9b1e-a71b705623cf req-cd99bbca-570d-4ce1-a492-ac656cedecc3 service nova] Releasing lock "refresh_cache-5854d21a-d1a8-4043-aec8-b37ff25c40e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.980077] env[65121]: DEBUG nova.network.neutron [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Successfully created port: 309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 725.103237] env[65121]: DEBUG nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 725.177803] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106232, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.375601] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e351091-9f43-48f0-a790-f47efdf670e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.380321] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 725.380321] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8437b3e1-1e8a-4452-b0e3-ab10b19066f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.388251] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221a76bf-d587-4c1f-acc3-54032a0e638b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.396697] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 725.396697] env[65121]: value = "task-5106234" [ 725.396697] env[65121]: _type = "Task" [ 725.396697] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.437915] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1142151-cf10-4a60-baab-ba892514053e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.444201] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106234, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.454710] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe9d267-b3f1-4cb9-b0a6-0f8f20020733 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.459443] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106233, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.472707] env[65121]: DEBUG nova.compute.provider_tree [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.671391] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106232, 'name': CreateVM_Task, 'duration_secs': 1.312036} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.671574] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1004635-b318-489c-9e16-6cb545279953] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 725.672113] env[65121]: WARNING neutronclient.v2_0.client [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.672654] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.672799] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.673123] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 725.676398] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21bcf507-76a5-4b81-8550-7b5a1c2a3840 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.680334] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 725.680334] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5267b6e6-8037-7a7e-3bfc-29aebfb39a8c" [ 725.680334] env[65121]: _type = "Task" [ 725.680334] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.690134] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5267b6e6-8037-7a7e-3bfc-29aebfb39a8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.762566] env[65121]: DEBUG oslo_concurrency.lockutils [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Acquiring lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.762892] env[65121]: DEBUG oslo_concurrency.lockutils [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.763145] env[65121]: INFO nova.compute.manager [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Attaching volume 47f77546-e67e-431f-9d50-fad84e2418eb to /dev/sdb [ 725.820880] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91360e23-ba87-49c6-ba1b-4b84ca94ffe4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.836209] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d074c059-00eb-462e-bade-5a1a03d5b4d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.855152] env[65121]: DEBUG nova.virt.block_device [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Updating existing volume attachment record: 1813260b-7d83-42c6-bdd4-4d02c68c6f58 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 725.909431] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106234, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.943240] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106233, 'name': ReconfigVM_Task, 'duration_secs': 1.089689} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.948149] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f/1e4a9d58-f848-4ff2-8558-4f23c6082a5f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 725.948149] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3822c4a-4324-4771-9542-9c63b44324da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.955486] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 725.955486] env[65121]: value = "task-5106235" [ 725.955486] env[65121]: _type = "Task" [ 725.955486] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.968702] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106235, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.975855] env[65121]: DEBUG nova.scheduler.client.report [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 726.117457] env[65121]: DEBUG nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 726.157575] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 726.158187] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 726.158187] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 726.158187] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 726.158680] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 726.158680] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 726.158680] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.158864] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 726.158983] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 726.159129] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 726.159262] env[65121]: DEBUG nova.virt.hardware [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 726.160283] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66556714-0665-4d87-a626-2f5285115f11 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.169806] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850db8eb-4666-41fd-a723-c866da6a4538 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.196749] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5267b6e6-8037-7a7e-3bfc-29aebfb39a8c, 'name': SearchDatastore_Task, 'duration_secs': 0.012992} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.197505] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.197776] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.198038] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.198184] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.198441] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.198789] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1bd63d4b-bf64-4dab-9071-51b11f20d18b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.211868] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.212039] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 726.212804] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c36d0a01-f582-45b2-ad49-0c0b99e9c250 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.220564] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 726.220564] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b02456-f6a6-f15c-dc12-b32682f515c6" [ 726.220564] env[65121]: _type = "Task" [ 726.220564] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.230294] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b02456-f6a6-f15c-dc12-b32682f515c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.286057] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.286353] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.286636] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.286862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.287112] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.290264] env[65121]: INFO nova.compute.manager [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Terminating instance [ 726.411195] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106234, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.473029] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106235, 'name': Rename_Task, 'duration_secs': 0.178381} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.473339] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 726.473777] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b23c20d-7e1c-43ae-bd9e-4df10469a19e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.482373] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.396s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.485659] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.806s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.485918] env[65121]: DEBUG nova.objects.instance [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lazy-loading 'resources' on Instance uuid 61bd5995-701e-430f-9aae-4b266089e313 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.487414] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 726.487414] env[65121]: value = "task-5106240" [ 726.487414] env[65121]: _type = "Task" [ 726.487414] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.498953] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106240, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.513605] env[65121]: INFO nova.scheduler.client.report [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Deleted allocations for instance 124876fe-a4f2-4140-8dee-d8bb72b5c9a5 [ 726.737532] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b02456-f6a6-f15c-dc12-b32682f515c6, 'name': SearchDatastore_Task, 'duration_secs': 0.021822} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.737971] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9d007a8-3e96-4828-9ab9-4c63802c77de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.744781] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 726.744781] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527d0a19-c891-49ef-588f-a6f577ac83b5" [ 726.744781] env[65121]: _type = "Task" [ 726.744781] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.755394] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527d0a19-c891-49ef-588f-a6f577ac83b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.781512] env[65121]: DEBUG nova.compute.manager [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Received event network-changed-2cd8c718-cf4c-4d35-92ae-d95f45242770 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 726.781695] env[65121]: DEBUG nova.compute.manager [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Refreshing instance network info cache due to event network-changed-2cd8c718-cf4c-4d35-92ae-d95f45242770. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 726.781898] env[65121]: DEBUG oslo_concurrency.lockutils [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Acquiring lock "refresh_cache-c1004635-b318-489c-9e16-6cb545279953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.782046] env[65121]: DEBUG oslo_concurrency.lockutils [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Acquired lock "refresh_cache-c1004635-b318-489c-9e16-6cb545279953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.782197] env[65121]: DEBUG nova.network.neutron [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Refreshing network info cache for port 2cd8c718-cf4c-4d35-92ae-d95f45242770 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 726.794368] env[65121]: DEBUG nova.compute.manager [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 726.794853] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.796044] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bb2b56-edec-4e3e-844f-8a20c7990437 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.805962] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 726.806332] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e513290d-eb32-4072-a54e-d6d146c9bda7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.819275] env[65121]: DEBUG oslo_vmware.api [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 726.819275] env[65121]: value = "task-5106241" [ 726.819275] env[65121]: _type = "Task" [ 726.819275] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.840056] env[65121]: DEBUG oslo_vmware.api [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106241, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.841988] env[65121]: DEBUG nova.network.neutron [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Successfully updated port: 309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 726.853070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.853753] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.853970] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.854491] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.854802] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.859327] env[65121]: INFO nova.compute.manager [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Terminating instance [ 726.914549] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106234, 'name': CreateSnapshot_Task, 'duration_secs': 1.095169} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.916114] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 726.916340] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5067c12-79d3-421d-a73b-a3acfceae1ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.004789] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106240, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.026148] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.026494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.027768] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca4c6cd8-49cc-423a-8bf1-3ad3421fe33b tempest-DeleteServersAdminTestJSON-15931195 tempest-DeleteServersAdminTestJSON-15931195-project-member] Lock "124876fe-a4f2-4140-8dee-d8bb72b5c9a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.308s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.258046] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527d0a19-c891-49ef-588f-a6f577ac83b5, 'name': SearchDatastore_Task, 'duration_secs': 0.013683} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.258353] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.258563] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c1004635-b318-489c-9e16-6cb545279953/c1004635-b318-489c-9e16-6cb545279953.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 727.259753] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a49f8fc-2777-415e-9ee6-4b06beac4301 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.269667] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 727.269667] env[65121]: value = "task-5106242" [ 727.269667] env[65121]: _type = "Task" [ 727.269667] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.278682] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106242, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.285424] env[65121]: WARNING neutronclient.v2_0.client [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.286824] env[65121]: WARNING openstack [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.286824] env[65121]: WARNING openstack [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.333855] env[65121]: DEBUG oslo_vmware.api [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106241, 'name': PowerOffVM_Task, 'duration_secs': 0.203077} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.336273] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 727.336435] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 727.336882] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0574fb12-f004-4b0c-bb8c-cdf92d7f5ac7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.346780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.346780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquired lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.346780] env[65121]: DEBUG nova.network.neutron [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 727.367679] env[65121]: DEBUG nova.compute.manager [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 727.367679] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.368825] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836c1121-b664-4026-aa95-17ae64e78bf3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.379868] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 727.379868] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88f711be-5b77-4f6d-9cfb-5e0d5c15d8db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.391897] env[65121]: DEBUG oslo_vmware.api [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 727.391897] env[65121]: value = "task-5106244" [ 727.391897] env[65121]: _type = "Task" [ 727.391897] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.411109] env[65121]: DEBUG oslo_vmware.api [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.433308] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 727.434156] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 727.434156] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Deleting the datastore file [datastore1] 5854d21a-d1a8-4043-aec8-b37ff25c40e7 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 727.442150] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 727.453624] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4b4904d-3e1a-405c-972a-f172bb7207ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.455485] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-359e8462-09b7-4424-869d-b19f07fe8f3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.471097] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 727.471097] env[65121]: value = "task-5106246" [ 727.471097] env[65121]: _type = "Task" [ 727.471097] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.471912] env[65121]: DEBUG oslo_vmware.api [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for the task: (returnval){ [ 727.471912] env[65121]: value = "task-5106245" [ 727.471912] env[65121]: _type = "Task" [ 727.471912] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.481708] env[65121]: WARNING openstack [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.482151] env[65121]: WARNING openstack [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.505116] env[65121]: DEBUG oslo_vmware.api [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.505480] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106246, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.513169] env[65121]: DEBUG oslo_vmware.api [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106240, 'name': PowerOnVM_Task, 'duration_secs': 0.612704} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.513594] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 727.513822] env[65121]: DEBUG nova.compute.manager [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 727.514781] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944b2ee0-f13c-4281-9d2e-43f389c0afa3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.601014] env[65121]: WARNING neutronclient.v2_0.client [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.601932] env[65121]: WARNING openstack [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.602436] env[65121]: WARNING openstack [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.713448] env[65121]: DEBUG nova.network.neutron [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Updated VIF entry in instance network info cache for port 2cd8c718-cf4c-4d35-92ae-d95f45242770. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 727.713834] env[65121]: DEBUG nova.network.neutron [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Updating instance_info_cache with network_info: [{"id": "2cd8c718-cf4c-4d35-92ae-d95f45242770", "address": "fa:16:3e:ff:2e:63", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd8c718-cf", "ovs_interfaceid": "2cd8c718-cf4c-4d35-92ae-d95f45242770", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 727.765949] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ff5d76-3064-4bcb-a372-b6dd7f9d6a86 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.784309] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2507e1a-6ee8-4fb3-88c0-a3d99bd36fcf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.787966] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106242, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.821815] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "6164d64c-75c5-469e-ac32-667842b3717a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.822116] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "6164d64c-75c5-469e-ac32-667842b3717a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.822327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "6164d64c-75c5-469e-ac32-667842b3717a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.822519] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "6164d64c-75c5-469e-ac32-667842b3717a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.822681] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "6164d64c-75c5-469e-ac32-667842b3717a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.825270] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dac7f45-4cca-46b3-9754-37dbadd2cff0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.828952] env[65121]: INFO nova.compute.manager [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Terminating instance [ 727.838869] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef92fe64-63c6-489d-9359-fbddfe412885 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.859727] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.860854] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.868595] env[65121]: DEBUG nova.compute.provider_tree [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.906285] env[65121]: DEBUG oslo_vmware.api [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106244, 'name': PowerOffVM_Task, 'duration_secs': 0.488899} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.906770] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 727.907054] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 727.907341] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d451f7ff-5edc-42d7-afc6-f76d405388a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.913498] env[65121]: DEBUG nova.network.neutron [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 727.970286] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.970286] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.993954] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 727.994350] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 727.994953] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleting the datastore file [datastore2] ab27fa56-f672-4096-a8f4-cce5ff4d5460 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 728.003146] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06d6f629-aafb-4a28-99e3-1bc162af2811 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.005461] env[65121]: DEBUG oslo_vmware.api [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Task: {'id': task-5106245, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.471408} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.005746] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106246, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.006184] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.007535] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 728.007535] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.007535] env[65121]: INFO nova.compute.manager [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Took 1.21 seconds to destroy the instance on the hypervisor. [ 728.007535] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 728.007778] env[65121]: DEBUG nova.compute.manager [-] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 728.008126] env[65121]: DEBUG nova.network.neutron [-] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 728.008406] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.009030] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.010367] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.020757] env[65121]: DEBUG oslo_vmware.api [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 728.020757] env[65121]: value = "task-5106248" [ 728.020757] env[65121]: _type = "Task" [ 728.020757] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.036878] env[65121]: DEBUG oslo_vmware.api [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106248, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.039139] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.099410] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "c1632a96-f334-4978-9f31-97f9329577e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.099410] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "c1632a96-f334-4978-9f31-97f9329577e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.142403] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.158665] env[65121]: WARNING neutronclient.v2_0.client [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.159407] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.159821] env[65121]: WARNING openstack [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.220470] env[65121]: DEBUG oslo_concurrency.lockutils [req-d35d096c-53aa-47c1-bb6c-5f3d0d9b6995 req-18686471-8f23-4310-9fbc-4ba459893a5b service nova] Releasing lock "refresh_cache-c1004635-b318-489c-9e16-6cb545279953" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.282872] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106242, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.671123} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.283301] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c1004635-b318-489c-9e16-6cb545279953/c1004635-b318-489c-9e16-6cb545279953.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 728.284184] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.284184] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-077f4ad0-ae66-4fca-b9ce-8841c349d0cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.295973] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 728.295973] env[65121]: value = "task-5106250" [ 728.295973] env[65121]: _type = "Task" [ 728.295973] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.306932] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.338165] env[65121]: DEBUG nova.compute.manager [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 728.338617] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.344984] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962e804d-ab72-43c7-8ff4-f2e26bb1dfa8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.354897] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 728.355517] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-406c4b9c-2b76-4e2a-9fba-5051df3024f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.364485] env[65121]: DEBUG oslo_vmware.api [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 728.364485] env[65121]: value = "task-5106251" [ 728.364485] env[65121]: _type = "Task" [ 728.364485] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.375225] env[65121]: DEBUG nova.scheduler.client.report [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 728.380691] env[65121]: DEBUG oslo_vmware.api [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.426354] env[65121]: DEBUG nova.network.neutron [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updating instance_info_cache with network_info: [{"id": "309f55e1-46cc-4806-9645-15db4d937a0b", "address": "fa:16:3e:29:ad:aa", "network": {"id": "08c3e4c7-f714-42eb-af35-74d1d1f92501", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1303322131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "0f0e2921e8704997ba4a1f5cf982989f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309f55e1-46", "ovs_interfaceid": "309f55e1-46cc-4806-9645-15db4d937a0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 728.496122] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106246, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.534162] env[65121]: DEBUG oslo_vmware.api [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106248, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301963} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.536303] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.536303] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 728.536303] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.536303] env[65121]: INFO nova.compute.manager [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Took 1.17 seconds to destroy the instance on the hypervisor. [ 728.536303] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 728.536303] env[65121]: DEBUG nova.compute.manager [-] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 728.536303] env[65121]: DEBUG nova.network.neutron [-] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 728.536303] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.536697] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.536892] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 728.592028] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.807927] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079382} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.807927] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 728.807927] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5686bae9-82f2-429d-bb4e-509aa176e852 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.838849] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] c1004635-b318-489c-9e16-6cb545279953/c1004635-b318-489c-9e16-6cb545279953.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 728.839727] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60543ec4-9730-4abb-b1e6-5b8cdabffc32 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.866982] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 728.866982] env[65121]: value = "task-5106252" [ 728.866982] env[65121]: _type = "Task" [ 728.866982] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.880846] env[65121]: DEBUG oslo_vmware.api [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106251, 'name': PowerOffVM_Task, 'duration_secs': 0.262241} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.881389] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106252, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.881892] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.396s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.884287] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 728.884488] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 728.885255] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.583s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.887362] env[65121]: INFO nova.compute.claims [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.894666] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5675c8a4-3f0d-42c8-80ce-006c86f842c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.919607] env[65121]: INFO nova.scheduler.client.report [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Deleted allocations for instance 61bd5995-701e-430f-9aae-4b266089e313 [ 728.931802] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Releasing lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.931802] env[65121]: DEBUG nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Instance network_info: |[{"id": "309f55e1-46cc-4806-9645-15db4d937a0b", "address": "fa:16:3e:29:ad:aa", "network": {"id": "08c3e4c7-f714-42eb-af35-74d1d1f92501", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1303322131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "0f0e2921e8704997ba4a1f5cf982989f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309f55e1-46", "ovs_interfaceid": "309f55e1-46cc-4806-9645-15db4d937a0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 728.933062] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:ad:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06cc7c49-c46c-4c1e-bf51-77e9ea802c40', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '309f55e1-46cc-4806-9645-15db4d937a0b', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 728.945873] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Creating folder: Project (0f0e2921e8704997ba4a1f5cf982989f). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 728.947220] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2c9e80f-6e1e-4528-90a4-9243686c3184 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.962033] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Created folder: Project (0f0e2921e8704997ba4a1f5cf982989f) in parent group-v993268. [ 728.962176] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Creating folder: Instances. Parent ref: group-v993372. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 728.962436] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3da54e4f-c46c-4cd4-a97f-4335837ab061 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.980780] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Created folder: Instances in parent group-v993372. [ 728.980780] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 728.980780] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 728.980982] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 728.981375] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleting the datastore file [datastore2] 6164d64c-75c5-469e-ac32-667842b3717a {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 728.981499] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 728.982307] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9dc2059f-4544-494f-ac5e-c8ae2cafbac0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.984748] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2e5342c-8bb2-4c25-b3f3-2e9d92c8de1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.012802] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106246, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.016105] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.016105] env[65121]: value = "task-5106257" [ 729.016105] env[65121]: _type = "Task" [ 729.016105] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.016105] env[65121]: DEBUG oslo_vmware.api [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for the task: (returnval){ [ 729.016105] env[65121]: value = "task-5106256" [ 729.016105] env[65121]: _type = "Task" [ 729.016105] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.030401] env[65121]: DEBUG oslo_vmware.api [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106256, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.034589] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106257, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.292089] env[65121]: DEBUG nova.network.neutron [-] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 729.302313] env[65121]: DEBUG nova.compute.manager [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Received event network-vif-plugged-309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 729.302560] env[65121]: DEBUG oslo_concurrency.lockutils [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Acquiring lock "06a99a15-5c8a-4be0-b393-80a104b9ba76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.302742] env[65121]: DEBUG oslo_concurrency.lockutils [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.303200] env[65121]: DEBUG oslo_concurrency.lockutils [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.303266] env[65121]: DEBUG nova.compute.manager [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] No waiting events found dispatching network-vif-plugged-309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 729.303498] env[65121]: WARNING nova.compute.manager [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Received unexpected event network-vif-plugged-309f55e1-46cc-4806-9645-15db4d937a0b for instance with vm_state building and task_state spawning. [ 729.305252] env[65121]: DEBUG nova.compute.manager [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Received event network-changed-309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 729.305252] env[65121]: DEBUG nova.compute.manager [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Refreshing instance network info cache due to event network-changed-309f55e1-46cc-4806-9645-15db4d937a0b. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 729.305252] env[65121]: DEBUG oslo_concurrency.lockutils [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Acquiring lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.305252] env[65121]: DEBUG oslo_concurrency.lockutils [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Acquired lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.305252] env[65121]: DEBUG nova.network.neutron [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Refreshing network info cache for port 309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 729.313863] env[65121]: DEBUG nova.network.neutron [-] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 729.332429] env[65121]: DEBUG nova.compute.manager [req-56bde0fe-a9bd-483d-a8e2-51e7f707f00a req-e44cdca9-987f-461c-adcf-349407546be9 service nova] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Received event network-vif-deleted-38e0b295-fa6e-4faa-be7b-8ec82a785df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 729.376785] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106252, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.431506] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.431763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.432283] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.432283] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.432390] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.438701] env[65121]: INFO nova.compute.manager [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Terminating instance [ 729.438701] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8141bad4-ab2f-45e1-a1c8-760290b6147a tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.647s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.439622] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 25.600s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.439829] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "61bd5995-701e-430f-9aae-4b266089e313-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.440026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.440187] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.442933] env[65121]: INFO nova.compute.manager [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Terminating instance [ 729.494267] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106246, 'name': CloneVM_Task, 'duration_secs': 1.778967} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.494679] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Created linked-clone VM from snapshot [ 729.495791] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9225a6e5-a080-43af-80ec-dc365a5b478e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.508756] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Uploading image 8480a96c-0246-4ea4-bd11-5ba473dc27b5 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 729.531381] env[65121]: DEBUG oslo_vmware.api [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Task: {'id': task-5106256, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198583} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.534196] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 729.534408] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 729.534592] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.534758] env[65121]: INFO nova.compute.manager [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Took 1.20 seconds to destroy the instance on the hypervisor. [ 729.535014] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 729.535555] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106257, 'name': CreateVM_Task, 'duration_secs': 0.393893} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.537596] env[65121]: DEBUG nova.compute.manager [-] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 729.537733] env[65121]: DEBUG nova.network.neutron [-] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 729.537973] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.538543] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.538799] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.545024] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 729.545867] env[65121]: WARNING neutronclient.v2_0.client [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.546362] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.546518] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.546893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 729.547451] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d8ee53a-2aaf-4b4c-aebb-ac959a35e278 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.552672] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 729.552672] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52be761e-f77e-107b-70ba-309d16c07af5" [ 729.552672] env[65121]: _type = "Task" [ 729.552672] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.558543] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 729.558543] env[65121]: value = "vm-993371" [ 729.558543] env[65121]: _type = "VirtualMachine" [ 729.558543] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 729.558825] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-214b606b-ad5d-426c-86f1-cc8bba38b7ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.566807] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be761e-f77e-107b-70ba-309d16c07af5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.569044] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lease: (returnval){ [ 729.569044] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523cef3b-034b-d8e9-8f81-f021f6c5991b" [ 729.569044] env[65121]: _type = "HttpNfcLease" [ 729.569044] env[65121]: } obtained for exporting VM: (result){ [ 729.569044] env[65121]: value = "vm-993371" [ 729.569044] env[65121]: _type = "VirtualMachine" [ 729.569044] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 729.569533] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the lease: (returnval){ [ 729.569533] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523cef3b-034b-d8e9-8f81-f021f6c5991b" [ 729.569533] env[65121]: _type = "HttpNfcLease" [ 729.569533] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 729.577251] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 729.577251] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523cef3b-034b-d8e9-8f81-f021f6c5991b" [ 729.577251] env[65121]: _type = "HttpNfcLease" [ 729.577251] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 729.582101] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.795480] env[65121]: INFO nova.compute.manager [-] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Took 1.26 seconds to deallocate network for instance. [ 729.807661] env[65121]: WARNING neutronclient.v2_0.client [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.808515] env[65121]: WARNING openstack [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.808903] env[65121]: WARNING openstack [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.819836] env[65121]: INFO nova.compute.manager [-] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Took 1.81 seconds to deallocate network for instance. [ 729.878756] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106252, 'name': ReconfigVM_Task, 'duration_secs': 0.781425} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.880714] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Reconfigured VM instance instance-00000023 to attach disk [datastore2] c1004635-b318-489c-9e16-6cb545279953/c1004635-b318-489c-9e16-6cb545279953.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 729.880714] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3846746e-8dfe-4293-9f02-5186bfd06192 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.889809] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 729.889809] env[65121]: value = "task-5106259" [ 729.889809] env[65121]: _type = "Task" [ 729.889809] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.904582] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106259, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.948717] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "refresh_cache-1e4a9d58-f848-4ff2-8558-4f23c6082a5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.948877] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquired lock "refresh_cache-1e4a9d58-f848-4ff2-8558-4f23c6082a5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.949048] env[65121]: DEBUG nova.network.neutron [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 729.951148] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.951313] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquired lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.951470] env[65121]: DEBUG nova.network.neutron [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 730.067471] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be761e-f77e-107b-70ba-309d16c07af5, 'name': SearchDatastore_Task, 'duration_secs': 0.015206} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.067803] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.067996] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 730.068779] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.068951] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.069152] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 730.070353] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4692332-1d06-46b0-9d60-35fce896b0a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.082437] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 730.082437] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523cef3b-034b-d8e9-8f81-f021f6c5991b" [ 730.082437] env[65121]: _type = "HttpNfcLease" [ 730.082437] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 730.087611] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 730.087611] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523cef3b-034b-d8e9-8f81-f021f6c5991b" [ 730.087611] env[65121]: _type = "HttpNfcLease" [ 730.087611] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 730.091814] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c989c366-d821-4dd3-8f35-ede3fa853d5d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.094311] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 730.094582] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 730.097820] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-024f4774-41f1-4f02-ab45-3d93e9d82ebb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.106595] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 730.106595] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523ba6b7-a30a-0af9-db19-3dc0c41f8738" [ 730.106595] env[65121]: _type = "Task" [ 730.106595] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.110410] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528dcce5-ac3d-ae13-3197-3ecb653fbe76/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 730.110599] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528dcce5-ac3d-ae13-3197-3ecb653fbe76/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 730.180709] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523ba6b7-a30a-0af9-db19-3dc0c41f8738, 'name': SearchDatastore_Task, 'duration_secs': 0.01061} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.183466] env[65121]: WARNING openstack [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.186876] env[65121]: WARNING openstack [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.193962] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd7d7581-e75f-478a-ac64-ed97b9e1da5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.203186] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 730.203186] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523f284a-1c96-b422-6e96-bbce12ef9f2c" [ 730.203186] env[65121]: _type = "Task" [ 730.203186] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.215751] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523f284a-1c96-b422-6e96-bbce12ef9f2c, 'name': SearchDatastore_Task, 'duration_secs': 0.011051} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.216439] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.216439] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 06a99a15-5c8a-4be0-b393-80a104b9ba76/06a99a15-5c8a-4be0-b393-80a104b9ba76.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 730.216727] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-afb0fa41-86d9-44ce-a45a-776010b7826d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.228183] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a682fbf9-acde-4dcb-94f2-5172ec39e334 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.232449] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 730.232449] env[65121]: value = "task-5106260" [ 730.232449] env[65121]: _type = "Task" [ 730.232449] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.246990] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.300765] env[65121]: WARNING neutronclient.v2_0.client [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.302347] env[65121]: WARNING openstack [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.302347] env[65121]: WARNING openstack [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.312524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.326907] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.405017] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106259, 'name': Rename_Task, 'duration_secs': 0.213648} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.405162] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 730.406246] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62927395-c70c-4e68-9745-58da6112dbe6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.407942] env[65121]: DEBUG nova.network.neutron [-] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.421535] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 730.421535] env[65121]: value = "task-5106261" [ 730.421535] env[65121]: _type = "Task" [ 730.421535] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.439899] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106261, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.453933] env[65121]: WARNING neutronclient.v2_0.client [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.456639] env[65121]: WARNING openstack [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.456639] env[65121]: WARNING openstack [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.469969] env[65121]: DEBUG nova.compute.utils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Can not refresh info_cache because instance was not found {{(pid=65121) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 730.471100] env[65121]: WARNING neutronclient.v2_0.client [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.471443] env[65121]: WARNING openstack [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.475132] env[65121]: WARNING openstack [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.490890] env[65121]: DEBUG nova.network.neutron [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updated VIF entry in instance network info cache for port 309f55e1-46cc-4806-9645-15db4d937a0b. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 730.491500] env[65121]: DEBUG nova.network.neutron [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updating instance_info_cache with network_info: [{"id": "309f55e1-46cc-4806-9645-15db4d937a0b", "address": "fa:16:3e:29:ad:aa", "network": {"id": "08c3e4c7-f714-42eb-af35-74d1d1f92501", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1303322131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "0f0e2921e8704997ba4a1f5cf982989f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309f55e1-46", "ovs_interfaceid": "309f55e1-46cc-4806-9645-15db4d937a0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.510936] env[65121]: DEBUG nova.network.neutron [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 730.535394] env[65121]: DEBUG nova.network.neutron [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 730.667650] env[65121]: DEBUG nova.network.neutron [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.745614] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106260, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.811637] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70c77e1-13f6-4576-82e2-3366e092aee9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.823905] env[65121]: DEBUG nova.network.neutron [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.827226] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb9d78c-009e-4e18-9902-e36f5ec279ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.865114] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662c5370-6db0-47d4-a6f5-1efd4daf5def {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.874453] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9547466-840f-4fcc-9b18-c02661fa5484 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.891417] env[65121]: DEBUG nova.compute.provider_tree [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.910657] env[65121]: INFO nova.compute.manager [-] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Took 1.37 seconds to deallocate network for instance. [ 730.933344] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106261, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.951863] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 730.952152] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993369', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'name': 'volume-47f77546-e67e-431f-9d50-fad84e2418eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818', 'attached_at': '', 'detached_at': '', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'serial': '47f77546-e67e-431f-9d50-fad84e2418eb'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 730.953031] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899e6a0f-ef56-4557-81a6-36133c85dbc0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.974344] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088c8023-0e60-41f3-be03-de536ae707b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.999890] env[65121]: DEBUG oslo_concurrency.lockutils [req-71fd94e5-0f5e-44c1-80ba-dff565bbebc9 req-a57d8ef8-4584-4fe5-a228-6e1d3e2f2a72 service nova] Releasing lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.009081] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] volume-47f77546-e67e-431f-9d50-fad84e2418eb/volume-47f77546-e67e-431f-9d50-fad84e2418eb.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.010642] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33a0a720-b53e-4ab8-9840-2689555285dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.034476] env[65121]: DEBUG oslo_vmware.api [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Waiting for the task: (returnval){ [ 731.034476] env[65121]: value = "task-5106262" [ 731.034476] env[65121]: _type = "Task" [ 731.034476] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.048270] env[65121]: DEBUG oslo_vmware.api [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106262, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.175560] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Releasing lock "refresh_cache-1e4a9d58-f848-4ff2-8558-4f23c6082a5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.176084] env[65121]: DEBUG nova.compute.manager [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 731.176238] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.177153] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1606e654-331b-4942-b543-af83a3f971f2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.187295] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 731.188462] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-699b9b93-57bc-4bd9-98a0-ee004c09ddbd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.197311] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 731.197311] env[65121]: value = "task-5106263" [ 731.197311] env[65121]: _type = "Task" [ 731.197311] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.210542] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.246954] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106260, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.327510] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Releasing lock "refresh_cache-61bd5995-701e-430f-9aae-4b266089e313" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.327945] env[65121]: DEBUG nova.compute.manager [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 731.328211] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.328545] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dde097a2-d6a3-462d-b534-df37e95c1627 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.342245] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61f9fb8-2f53-4e9d-9d59-cbc74bdf2d75 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.389183] env[65121]: WARNING nova.virt.vmwareapi.vmops [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61bd5995-701e-430f-9aae-4b266089e313 could not be found. [ 731.389429] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 731.389642] env[65121]: INFO nova.compute.manager [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Took 0.06 seconds to destroy the instance on the hypervisor. [ 731.389903] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 731.390262] env[65121]: DEBUG nova.compute.manager [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 731.390360] env[65121]: DEBUG nova.network.neutron [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 731.390607] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 731.391170] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.391433] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.399981] env[65121]: DEBUG nova.scheduler.client.report [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 731.418601] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.438422] env[65121]: DEBUG oslo_vmware.api [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106261, 'name': PowerOnVM_Task, 'duration_secs': 0.74858} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.438422] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 731.438660] env[65121]: INFO nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Took 10.47 seconds to spawn the instance on the hypervisor. [ 731.438706] env[65121]: DEBUG nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 731.439614] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f1bfd7-0fb5-49de-8f67-1752e165dd98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.548849] env[65121]: DEBUG oslo_vmware.api [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106262, 'name': ReconfigVM_Task, 'duration_secs': 0.435516} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.549545] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Reconfigured VM instance instance-0000000e to attach disk [datastore2] volume-47f77546-e67e-431f-9d50-fad84e2418eb/volume-47f77546-e67e-431f-9d50-fad84e2418eb.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.555218] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e675a93d-6e1a-450d-b0fe-de3bb40efbc7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.573567] env[65121]: DEBUG oslo_vmware.api [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Waiting for the task: (returnval){ [ 731.573567] env[65121]: value = "task-5106264" [ 731.573567] env[65121]: _type = "Task" [ 731.573567] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.584192] env[65121]: DEBUG oslo_vmware.api [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106264, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.708320] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.755607] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106260, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.791771] env[65121]: DEBUG nova.network.neutron [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 731.792058] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 731.905727] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.020s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.906560] env[65121]: DEBUG nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 731.909844] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 25.549s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.961583] env[65121]: INFO nova.compute.manager [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Took 47.19 seconds to build instance. [ 732.087105] env[65121]: DEBUG oslo_vmware.api [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106264, 'name': ReconfigVM_Task, 'duration_secs': 0.166983} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.087667] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993369', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'name': 'volume-47f77546-e67e-431f-9d50-fad84e2418eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818', 'attached_at': '', 'detached_at': '', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'serial': '47f77546-e67e-431f-9d50-fad84e2418eb'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 732.180694] env[65121]: DEBUG nova.compute.manager [req-8df80cd0-bb64-46a7-9357-3464d98f2916 req-c1d97720-62ad-4669-a606-25887f564291 service nova] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Received event network-vif-deleted-5604925f-4cea-418a-a42d-502bac3a6114 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 732.181235] env[65121]: DEBUG nova.compute.manager [req-8df80cd0-bb64-46a7-9357-3464d98f2916 req-c1d97720-62ad-4669-a606-25887f564291 service nova] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Received event network-vif-deleted-97194a81-8ff7-43dd-83f6-c4cce03f1985 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 732.209432] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.248569] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106260, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.54577} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.249061] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 06a99a15-5c8a-4be0-b393-80a104b9ba76/06a99a15-5c8a-4be0-b393-80a104b9ba76.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 732.249395] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.249687] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f73ac937-4dc1-46f3-8afe-0e3edb6c111a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.258375] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 732.258375] env[65121]: value = "task-5106265" [ 732.258375] env[65121]: _type = "Task" [ 732.258375] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.270506] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106265, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.300344] env[65121]: DEBUG nova.network.neutron [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 732.415846] env[65121]: DEBUG nova.compute.utils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 732.417915] env[65121]: DEBUG nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 732.417915] env[65121]: DEBUG nova.network.neutron [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 732.418395] env[65121]: WARNING neutronclient.v2_0.client [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.418817] env[65121]: WARNING neutronclient.v2_0.client [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.419531] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.419921] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.441119] env[65121]: DEBUG nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 732.465542] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1a72f0a0-4f56-4b79-b8fb-de1139b2d4ab tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c1004635-b318-489c-9e16-6cb545279953" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.695s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.496528] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "aa9b6708-c53c-4117-9b75-9d506f393395" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.496865] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "aa9b6708-c53c-4117-9b75-9d506f393395" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.497103] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "aa9b6708-c53c-4117-9b75-9d506f393395-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.497307] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "aa9b6708-c53c-4117-9b75-9d506f393395-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.497483] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "aa9b6708-c53c-4117-9b75-9d506f393395-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.500660] env[65121]: INFO nova.compute.manager [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Terminating instance [ 732.515767] env[65121]: DEBUG nova.policy [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e773a3343a448f5acd3f08005e8ed45', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '40aee06073b04c25bc7b38fbd09856e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 732.710128] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106263, 'name': PowerOffVM_Task, 'duration_secs': 1.226991} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.710128] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.710128] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.710128] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9a01a3b-3007-4aba-81e2-530743326c39 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.738971] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.739275] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.739657] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Deleting the datastore file [datastore2] 1e4a9d58-f848-4ff2-8558-4f23c6082a5f {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.739770] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86bca1ec-0efd-46d7-8cde-ad15d0be9034 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.750574] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for the task: (returnval){ [ 732.750574] env[65121]: value = "task-5106267" [ 732.750574] env[65121]: _type = "Task" [ 732.750574] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.759777] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106267, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.767994] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106265, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079223} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.768363] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.769183] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafed9f0-db8b-46ac-a0cc-d853e5a05590 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.795713] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 06a99a15-5c8a-4be0-b393-80a104b9ba76/06a99a15-5c8a-4be0-b393-80a104b9ba76.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.797988] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d1c4edc-5363-4f72-9084-6f8291dbf995 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.815338] env[65121]: INFO nova.compute.manager [-] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Took 1.42 seconds to deallocate network for instance. [ 732.823835] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 732.823835] env[65121]: value = "task-5106268" [ 732.823835] env[65121]: _type = "Task" [ 732.823835] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.833544] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.860476] env[65121]: DEBUG nova.network.neutron [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Successfully created port: b5cdcd85-1d9d-4eb2-8705-20970c1d609e {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 732.969433] env[65121]: DEBUG nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 732.982128] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance a43879c7-7378-4700-9581-8daabdafe5b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.982359] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance aa9b6708-c53c-4117-9b75-9d506f393395 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.982477] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance cc002961-d742-4255-88c7-f0a5a39424b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.982585] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f1a5fdce-a940-43d4-83d7-716786c9ac34 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.982689] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c4096314-270a-4270-9e1d-5ace8ddbd286 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.982791] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.982917] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 46f89cfb-14aa-483d-aa8a-bc91356272ca is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 732.983032] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 5f6e4f46-0745-42c0-a779-6cffb60e21ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.983147] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 5854d21a-d1a8-4043-aec8-b37ff25c40e7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 732.983249] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance e2d32a5e-c350-4b2b-9243-c3b412193a82 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 6a4841b6-6076-4e9c-a6cf-a658957e2931 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 732.985019] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 7a0d0769-31c5-4ae1-8520-a744f64d39c6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4b6fe349-335e-4202-ab76-04f6ccc036ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c993d5a4-49d6-43aa-a1f0-4aac91fad953 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance cd86f815-4b46-45b1-9a1c-ab2e28f63626 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c2f9fa49-4273-4885-b59d-0635ed739d65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 27940143-16b5-4263-b23c-354ed8ea8866 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 1e4a9d58-f848-4ff2-8558-4f23c6082a5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance b1ad9d15-7ef4-404a-9751-ecdd02ea768b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 9cf6c29a-3424-4b88-9ba5-8120b124beb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985019] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 6164d64c-75c5-469e-ac32-667842b3717a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 732.985443] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance ab27fa56-f672-4096-a8f4-cce5ff4d5460 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 732.985443] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c1004635-b318-489c-9e16-6cb545279953 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985443] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 06a99a15-5c8a-4be0-b393-80a104b9ba76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 732.985539] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 733.005724] env[65121]: DEBUG nova.compute.manager [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 733.005999] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.008153] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf749900-e2cd-4881-a452-5088d2f00e77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.020661] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 733.021070] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88a757aa-ada6-467b-80fc-70227183bd4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.031184] env[65121]: DEBUG oslo_vmware.api [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 733.031184] env[65121]: value = "task-5106269" [ 733.031184] env[65121]: _type = "Task" [ 733.031184] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.044185] env[65121]: DEBUG oslo_vmware.api [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106269, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.147035] env[65121]: DEBUG nova.objects.instance [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lazy-loading 'flavor' on Instance uuid 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 733.271257] env[65121]: DEBUG oslo_vmware.api [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Task: {'id': task-5106267, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149396} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.271257] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 733.271257] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 733.271257] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.271257] env[65121]: INFO nova.compute.manager [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Took 2.09 seconds to destroy the instance on the hypervisor. [ 733.271257] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 733.271257] env[65121]: DEBUG nova.compute.manager [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 733.271257] env[65121]: DEBUG nova.network.neutron [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 733.271257] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 733.272423] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.272861] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.305108] env[65121]: DEBUG nova.network.neutron [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 733.305532] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 733.322717] env[65121]: INFO nova.compute.manager [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance disappeared during terminate [ 733.322953] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9576e149-42e2-4293-ac55-78ae4ab8e027 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "61bd5995-701e-430f-9aae-4b266089e313" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.883s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.339808] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106268, 'name': ReconfigVM_Task, 'duration_secs': 0.436599} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.340055] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 06a99a15-5c8a-4be0-b393-80a104b9ba76/06a99a15-5c8a-4be0-b393-80a104b9ba76.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.340727] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-144595d3-d90b-4318-8b17-d472d0ee6724 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.350031] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 733.350031] env[65121]: value = "task-5106270" [ 733.350031] env[65121]: _type = "Task" [ 733.350031] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.363111] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106270, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.457470] env[65121]: DEBUG nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 733.483990] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 733.483990] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 733.484572] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 733.484572] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 733.484572] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 733.484700] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 733.484978] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.485200] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 733.485419] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 733.485610] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 733.485829] env[65121]: DEBUG nova.virt.hardware [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 733.487334] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb5c03f-8144-4f8e-a8f2-ad6b7e96a4d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.494440] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 50083ee5-9655-4cab-9d50-04a97baac626 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 733.497240] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.503339] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47c5987-15c2-46d6-b495-c050c82832ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.547489] env[65121]: DEBUG oslo_vmware.api [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106269, 'name': PowerOffVM_Task, 'duration_secs': 0.252908} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.547704] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 733.547981] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 733.548172] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-841f5066-e9fa-4806-8e38-0931b8a10e40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.627871] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 733.628855] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 733.628855] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Deleting the datastore file [datastore2] aa9b6708-c53c-4117-9b75-9d506f393395 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 733.628855] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df1ef56f-40a3-44e2-a440-446f5606690d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.639491] env[65121]: DEBUG oslo_vmware.api [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for the task: (returnval){ [ 733.639491] env[65121]: value = "task-5106272" [ 733.639491] env[65121]: _type = "Task" [ 733.639491] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.649431] env[65121]: DEBUG oslo_vmware.api [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106272, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.653520] env[65121]: DEBUG oslo_concurrency.lockutils [None req-21a4152d-9603-4880-9f77-af38cc0adfa3 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.891s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.807866] env[65121]: DEBUG nova.network.neutron [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 733.865360] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106270, 'name': Rename_Task, 'duration_secs': 0.205777} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.865670] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 733.865940] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca9f5ead-38f3-46d1-86de-204838477cc0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.874039] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 733.874039] env[65121]: value = "task-5106273" [ 733.874039] env[65121]: _type = "Task" [ 733.874039] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.889171] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106273, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.996791] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 3f69040e-7df3-4535-a3a7-90f3348ef346 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 734.160643] env[65121]: DEBUG oslo_vmware.api [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Task: {'id': task-5106272, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.225528} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.161275] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 734.161517] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 734.162041] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.162041] env[65121]: INFO nova.compute.manager [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Took 1.16 seconds to destroy the instance on the hypervisor. [ 734.162173] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 734.162307] env[65121]: DEBUG nova.compute.manager [-] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 734.162368] env[65121]: DEBUG nova.network.neutron [-] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 734.162640] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.163425] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.163948] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.312022] env[65121]: INFO nova.compute.manager [-] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Took 1.04 seconds to deallocate network for instance. [ 734.358602] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.387565] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106273, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.504752] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f84d7aa1-a90b-4c95-9673-fb1a2af7187c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 734.701208] env[65121]: DEBUG nova.network.neutron [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Successfully updated port: b5cdcd85-1d9d-4eb2-8705-20970c1d609e {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 734.819582] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.886492] env[65121]: DEBUG oslo_vmware.api [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106273, 'name': PowerOnVM_Task, 'duration_secs': 0.71422} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.886851] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 734.887097] env[65121]: INFO nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Took 8.77 seconds to spawn the instance on the hypervisor. [ 734.887265] env[65121]: DEBUG nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 734.888113] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f558f96-5b74-4b1c-849d-22efe849eb57 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.009505] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 41af3ee5-0ee6-4957-bad0-5aef51d55bd5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 735.184043] env[65121]: DEBUG nova.compute.manager [req-fdd71882-4007-41b7-85eb-504d92425215 req-2bfe8003-b50f-4331-8a0f-82230803a4b8 service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Received event network-vif-plugged-b5cdcd85-1d9d-4eb2-8705-20970c1d609e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 735.184274] env[65121]: DEBUG oslo_concurrency.lockutils [req-fdd71882-4007-41b7-85eb-504d92425215 req-2bfe8003-b50f-4331-8a0f-82230803a4b8 service nova] Acquiring lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.184572] env[65121]: DEBUG oslo_concurrency.lockutils [req-fdd71882-4007-41b7-85eb-504d92425215 req-2bfe8003-b50f-4331-8a0f-82230803a4b8 service nova] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.184759] env[65121]: DEBUG oslo_concurrency.lockutils [req-fdd71882-4007-41b7-85eb-504d92425215 req-2bfe8003-b50f-4331-8a0f-82230803a4b8 service nova] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.184972] env[65121]: DEBUG nova.compute.manager [req-fdd71882-4007-41b7-85eb-504d92425215 req-2bfe8003-b50f-4331-8a0f-82230803a4b8 service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] No waiting events found dispatching network-vif-plugged-b5cdcd85-1d9d-4eb2-8705-20970c1d609e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 735.185372] env[65121]: WARNING nova.compute.manager [req-fdd71882-4007-41b7-85eb-504d92425215 req-2bfe8003-b50f-4331-8a0f-82230803a4b8 service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Received unexpected event network-vif-plugged-b5cdcd85-1d9d-4eb2-8705-20970c1d609e for instance with vm_state building and task_state spawning. [ 735.203125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "refresh_cache-5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.203349] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "refresh_cache-5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 735.203514] env[65121]: DEBUG nova.network.neutron [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 735.284208] env[65121]: DEBUG nova.network.neutron [-] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 735.349245] env[65121]: DEBUG nova.compute.manager [req-6d578f0b-2ad7-4e7a-a4e0-4992a7f1f8dd req-6a52b692-8685-4c84-bf0e-5cf6f27bf2d1 service nova] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Received event network-vif-deleted-f119a70b-b122-4875-b1a3-5c4c7a234c29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 735.421272] env[65121]: INFO nova.compute.manager [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Took 46.34 seconds to build instance. [ 735.512853] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 293f93f2-c01d-42c8-b1a7-3056805c77de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 735.707101] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.707693] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.762144] env[65121]: DEBUG nova.network.neutron [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 735.790021] env[65121]: INFO nova.compute.manager [-] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Took 1.62 seconds to deallocate network for instance. [ 735.796755] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.797878] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.861180] env[65121]: WARNING neutronclient.v2_0.client [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.862027] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.862486] env[65121]: WARNING openstack [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.872083] env[65121]: DEBUG oslo_concurrency.lockutils [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Acquiring lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.872343] env[65121]: DEBUG oslo_concurrency.lockutils [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.923060] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a7669cae-e00d-4a45-b8ab-ddb7cd03025c tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.810s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.987995] env[65121]: DEBUG nova.network.neutron [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Updating instance_info_cache with network_info: [{"id": "b5cdcd85-1d9d-4eb2-8705-20970c1d609e", "address": "fa:16:3e:b9:ad:f2", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5cdcd85-1d", "ovs_interfaceid": "b5cdcd85-1d9d-4eb2-8705-20970c1d609e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 736.016502] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 19b1f3f9-842e-4150-8890-b0b22393c3af has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 736.296406] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.377040] env[65121]: INFO nova.compute.manager [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Detaching volume 47f77546-e67e-431f-9d50-fad84e2418eb [ 736.422358] env[65121]: INFO nova.virt.block_device [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Attempting to driver detach volume 47f77546-e67e-431f-9d50-fad84e2418eb from mountpoint /dev/sdb [ 736.422552] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 736.422731] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993369', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'name': 'volume-47f77546-e67e-431f-9d50-fad84e2418eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818', 'attached_at': '', 'detached_at': '', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'serial': '47f77546-e67e-431f-9d50-fad84e2418eb'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 736.424057] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75596fc-2ad7-4188-aa33-d1f12403e8fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.430266] env[65121]: DEBUG nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 736.454522] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e127fd7-111f-4e92-ad3f-fd82db7240d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.462966] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dbf8c5-b399-40b2-badc-496e6b2e4cef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.485618] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99773c62-eb28-40f6-b9c5-1fa42467cd9b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.492307] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "refresh_cache-5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.492717] env[65121]: DEBUG nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Instance network_info: |[{"id": "b5cdcd85-1d9d-4eb2-8705-20970c1d609e", "address": "fa:16:3e:b9:ad:f2", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5cdcd85-1d", "ovs_interfaceid": "b5cdcd85-1d9d-4eb2-8705-20970c1d609e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 736.504474] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:ad:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5cdcd85-1d9d-4eb2-8705-20970c1d609e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.515134] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 736.515836] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] The volume has not been displaced from its original location: [datastore2] volume-47f77546-e67e-431f-9d50-fad84e2418eb/volume-47f77546-e67e-431f-9d50-fad84e2418eb.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 736.521843] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Reconfiguring VM instance instance-0000000e to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 736.522228] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 736.522969] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance aab9317b-4ee6-48b3-905b-859a5996f33d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 736.524257] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48b14f70-5d9f-4f91-bf85-d0c38f8971a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.538862] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fcc70a9-86d4-45ca-914f-819d05495a0a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.561496] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Waiting for the task: (returnval){ [ 736.561496] env[65121]: value = "task-5106274" [ 736.561496] env[65121]: _type = "Task" [ 736.561496] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.562379] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 736.562379] env[65121]: value = "task-5106275" [ 736.562379] env[65121]: _type = "Task" [ 736.562379] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.574802] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106275, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.578481] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106274, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.876778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.877024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.972884] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.039975] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 75114f97-fe50-4624-9333-303e411529ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 737.079985] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106274, 'name': ReconfigVM_Task, 'duration_secs': 0.311418} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.084622] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Reconfigured VM instance instance-0000000e to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 737.091231] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106275, 'name': CreateVM_Task, 'duration_secs': 0.436764} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.091570] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a759ba66-9e0c-425f-9fff-3c7d5740d2f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.104272] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 737.105236] env[65121]: WARNING neutronclient.v2_0.client [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 737.105645] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.105754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.106177] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 737.107040] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c380b32-ebb9-401f-a527-7f57b8f61ad7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.112463] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Waiting for the task: (returnval){ [ 737.112463] env[65121]: value = "task-5106276" [ 737.112463] env[65121]: _type = "Task" [ 737.112463] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.113669] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 737.113669] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52405740-50bf-f8ea-745a-500c641c4b09" [ 737.113669] env[65121]: _type = "Task" [ 737.113669] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.128056] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106276, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.130539] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52405740-50bf-f8ea-745a-500c641c4b09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.544509] env[65121]: DEBUG nova.compute.manager [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Received event network-changed-b5cdcd85-1d9d-4eb2-8705-20970c1d609e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 737.544509] env[65121]: DEBUG nova.compute.manager [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Refreshing instance network info cache due to event network-changed-b5cdcd85-1d9d-4eb2-8705-20970c1d609e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 737.544509] env[65121]: DEBUG oslo_concurrency.lockutils [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Acquiring lock "refresh_cache-5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.544509] env[65121]: DEBUG oslo_concurrency.lockutils [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Acquired lock "refresh_cache-5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.544509] env[65121]: DEBUG nova.network.neutron [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Refreshing network info cache for port b5cdcd85-1d9d-4eb2-8705-20970c1d609e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 737.545867] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance b6f6871d-311c-4adb-824e-2907a12f4224 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 737.626537] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106276, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.630196] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52405740-50bf-f8ea-745a-500c641c4b09, 'name': SearchDatastore_Task, 'duration_secs': 0.016791} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.630488] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.630721] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 737.630947] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.631094] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.631267] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 737.631524] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-46ab29b0-d790-4122-ae63-5636bfff8af0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.647998] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 737.648212] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 737.648970] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88ea0701-c82c-401c-973e-319f17502e78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.656325] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 737.656325] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526fa64a-16e9-ae0e-f715-d0e3022b8614" [ 737.656325] env[65121]: _type = "Task" [ 737.656325] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.664273] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526fa64a-16e9-ae0e-f715-d0e3022b8614, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.048441] env[65121]: WARNING neutronclient.v2_0.client [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.049254] env[65121]: WARNING openstack [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.049515] env[65121]: WARNING openstack [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.060141] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 00362477-c89e-4f60-98a4-d4928081d55e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 738.128228] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106276, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.174021] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526fa64a-16e9-ae0e-f715-d0e3022b8614, 'name': SearchDatastore_Task, 'duration_secs': 0.038688} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.176923] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc7b65bc-302f-41d8-8fad-e262711d59cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.184550] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 738.184550] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d6efb0-513b-d4b6-6c60-f9a60c335730" [ 738.184550] env[65121]: _type = "Task" [ 738.184550] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.194259] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d6efb0-513b-d4b6-6c60-f9a60c335730, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.197931] env[65121]: WARNING openstack [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.198398] env[65121]: WARNING openstack [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.282067] env[65121]: WARNING neutronclient.v2_0.client [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.282923] env[65121]: WARNING openstack [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.283322] env[65121]: WARNING openstack [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.381401] env[65121]: DEBUG nova.network.neutron [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Updated VIF entry in instance network info cache for port b5cdcd85-1d9d-4eb2-8705-20970c1d609e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 738.381764] env[65121]: DEBUG nova.network.neutron [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Updating instance_info_cache with network_info: [{"id": "b5cdcd85-1d9d-4eb2-8705-20970c1d609e", "address": "fa:16:3e:b9:ad:f2", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5cdcd85-1d", "ovs_interfaceid": "b5cdcd85-1d9d-4eb2-8705-20970c1d609e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 738.555309] env[65121]: DEBUG nova.compute.manager [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Received event network-changed-309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 738.555856] env[65121]: DEBUG nova.compute.manager [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Refreshing instance network info cache due to event network-changed-309f55e1-46cc-4806-9645-15db4d937a0b. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 738.555856] env[65121]: DEBUG oslo_concurrency.lockutils [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Acquiring lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.555856] env[65121]: DEBUG oslo_concurrency.lockutils [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Acquired lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.556548] env[65121]: DEBUG nova.network.neutron [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Refreshing network info cache for port 309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 738.561305] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance ab41ae41-e69d-47fb-a31a-16748fd3b0f2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 738.629623] env[65121]: DEBUG oslo_vmware.api [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Task: {'id': task-5106276, 'name': ReconfigVM_Task, 'duration_secs': 1.173446} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.629839] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993369', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'name': 'volume-47f77546-e67e-431f-9d50-fad84e2418eb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818', 'attached_at': '', 'detached_at': '', 'volume_id': '47f77546-e67e-431f-9d50-fad84e2418eb', 'serial': '47f77546-e67e-431f-9d50-fad84e2418eb'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 738.699035] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d6efb0-513b-d4b6-6c60-f9a60c335730, 'name': SearchDatastore_Task, 'duration_secs': 0.034976} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.699035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 738.699294] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f/5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 738.699479] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5aca15bf-6201-430c-9283-8c1acb79a530 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.707610] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 738.707610] env[65121]: value = "task-5106277" [ 738.707610] env[65121]: _type = "Task" [ 738.707610] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.717222] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.884574] env[65121]: DEBUG oslo_concurrency.lockutils [req-bce4a5fc-640e-4f21-ae39-072ebe902bac req-67610480-08ee-4ebb-ac91-8b6ed1b309fb service nova] Releasing lock "refresh_cache-5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.061656] env[65121]: WARNING neutronclient.v2_0.client [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.062932] env[65121]: WARNING openstack [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.063649] env[65121]: WARNING openstack [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.077996] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c1632a96-f334-4978-9f31-97f9329577e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 739.079044] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 20 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 739.079044] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4480MB phys_disk=100GB used_disk=20GB total_vcpus=48 used_vcpus=20 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '20', 'num_vm_active': '18', 'num_task_None': '17', 'num_os_type_None': '20', 'num_proj_982f66705583488cb369f19160cc2ee5': '1', 'io_workload': '2', 'num_proj_018269ed0c644392aebdb56cf0523292': '1', 'num_proj_e92570886f544d9ea9fe0947e8e93d9c': '1', 'num_proj_35312a302644426f98f127e89a067e75': '1', 'num_proj_ce90017874734825b9fbdce06eb12a00': '1', 'num_proj_72237a0d762645588c41231b0a34a796': '1', 'num_proj_def20b65ac8341c1a60a747260589d5a': '1', 'num_proj_741e03af2f4d451d9ed77004b644deb8': '1', 'num_proj_ec5741daa3f3411bb74b3b965bbeef44': '1', 'num_proj_5724c072955c4cd3a5500dedf3b972a3': '2', 'num_task_image_uploading': '1', 'num_proj_3c0dbd078f6b42d5ad8eeb0a45cac9f5': '3', 'num_proj_083c4f2aa7bf4167b472db6bebb9d8bd': '2', 'num_proj_4467da3ed41245ddbc93fc865a8b7bdd': '1', 'num_task_deleting': '1', 'num_proj_0bc48543d199422b9c3224d5602ed961': '1', 'num_vm_building': '2', 'num_task_spawning': '1', 'num_proj_0f0e2921e8704997ba4a1f5cf982989f': '1', 'num_proj_40aee06073b04c25bc7b38fbd09856e9': '1'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 739.206690] env[65121]: DEBUG nova.objects.instance [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lazy-loading 'flavor' on Instance uuid 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 739.231450] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106277, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.241707] env[65121]: WARNING openstack [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.243215] env[65121]: WARNING openstack [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.309477] env[65121]: WARNING neutronclient.v2_0.client [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.309664] env[65121]: WARNING openstack [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.310103] env[65121]: WARNING openstack [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.423241] env[65121]: DEBUG nova.network.neutron [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updated VIF entry in instance network info cache for port 309f55e1-46cc-4806-9645-15db4d937a0b. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 739.423613] env[65121]: DEBUG nova.network.neutron [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updating instance_info_cache with network_info: [{"id": "309f55e1-46cc-4806-9645-15db4d937a0b", "address": "fa:16:3e:29:ad:aa", "network": {"id": "08c3e4c7-f714-42eb-af35-74d1d1f92501", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1303322131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "0f0e2921e8704997ba4a1f5cf982989f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309f55e1-46", "ovs_interfaceid": "309f55e1-46cc-4806-9645-15db4d937a0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 739.732577] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106277, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653084} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.737947] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f/5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 739.739179] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 739.739179] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f43e4b9-d4c8-44da-abd5-39f65a38b0f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.752240] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 739.752240] env[65121]: value = "task-5106278" [ 739.752240] env[65121]: _type = "Task" [ 739.752240] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.765315] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106278, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.800627] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108af89f-5ee1-42dd-9df2-5df47cce3c84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.811067] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e97f6c-b3d3-4da0-8c34-8fa0c879d77c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.844822] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82eeec77-5758-4ce3-84b2-c63db2766566 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.853035] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036e798c-a2a6-494f-8901-271416d53f5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.869872] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.931509] env[65121]: DEBUG oslo_concurrency.lockutils [req-c9efe3f7-c0b3-482d-8f9e-593849e8217f req-2a841a6d-c214-4306-8ef0-ae7431b59d8b service nova] Releasing lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.201731] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528dcce5-ac3d-ae13-3197-3ecb653fbe76/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 740.202827] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf2fc69-7566-42e6-9099-83b48fb01237 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.210340] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528dcce5-ac3d-ae13-3197-3ecb653fbe76/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 740.210788] env[65121]: ERROR oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528dcce5-ac3d-ae13-3197-3ecb653fbe76/disk-0.vmdk due to incomplete transfer. [ 740.210788] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-79962713-ddf4-4fd5-a211-71f3f098db15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.219609] env[65121]: DEBUG oslo_vmware.rw_handles [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528dcce5-ac3d-ae13-3197-3ecb653fbe76/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 740.219803] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Uploaded image 8480a96c-0246-4ea4-bd11-5ba473dc27b5 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 740.222084] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 740.223468] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-018b3b8f-ecbf-45f7-922e-839ecebd22ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.230778] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 740.230778] env[65121]: value = "task-5106279" [ 740.230778] env[65121]: _type = "Task" [ 740.230778] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.245357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-93e653ed-602e-4968-9c91-dae6fd6321c5 tempest-VolumesAssistedSnapshotsTest-2131460256 tempest-VolumesAssistedSnapshotsTest-2131460256-project-admin] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.373s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.247832] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106279, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.263287] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106278, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091409} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.264218] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.265207] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb42f11-e506-45c8-a059-60b9cf0cd7e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.291411] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f/5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.292177] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5aca9785-dc81-481b-8aa6-ba42a33881a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.316248] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 740.316248] env[65121]: value = "task-5106280" [ 740.316248] env[65121]: _type = "Task" [ 740.316248] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.326323] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106280, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.373321] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 740.616086] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "c2f9fa49-4273-4885-b59d-0635ed739d65" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 740.616086] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.616086] env[65121]: DEBUG nova.compute.manager [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 740.616086] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-accfc580-e854-4258-a34c-b4663d4a6b2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.628302] env[65121]: DEBUG nova.compute.manager [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 740.630127] env[65121]: DEBUG nova.objects.instance [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lazy-loading 'flavor' on Instance uuid c2f9fa49-4273-4885-b59d-0635ed739d65 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 740.746134] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106279, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.828906] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106280, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.879304] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 740.879584] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.970s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.879902] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.820s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.881503] env[65121]: INFO nova.compute.claims [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.242766] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106279, 'name': Destroy_Task, 'duration_secs': 0.849078} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.243127] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Destroyed the VM [ 741.243453] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 741.243702] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4da2f29f-61cd-4c38-9d8d-9762623398b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.252596] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 741.252596] env[65121]: value = "task-5106282" [ 741.252596] env[65121]: _type = "Task" [ 741.252596] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.261828] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106282, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.329942] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106280, 'name': ReconfigVM_Task, 'duration_secs': 0.666672} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.331068] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Reconfigured VM instance instance-00000025 to attach disk [datastore2] 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f/5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 741.331068] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea33fc6b-d74d-469c-866a-162ce53f704f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.340566] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 741.340566] env[65121]: value = "task-5106283" [ 741.340566] env[65121]: _type = "Task" [ 741.340566] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.356843] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106283, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.640357] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 741.640719] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-478cd007-1d6d-4116-9830-bf4a0c7ade42 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.650971] env[65121]: DEBUG oslo_vmware.api [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 741.650971] env[65121]: value = "task-5106284" [ 741.650971] env[65121]: _type = "Task" [ 741.650971] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.665033] env[65121]: DEBUG oslo_vmware.api [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106284, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.764199] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106282, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.852152] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106283, 'name': Rename_Task, 'duration_secs': 0.160315} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.852463] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 741.852715] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e187f9ae-78f7-4e02-ab4a-c2270765f26a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.862696] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 741.862696] env[65121]: value = "task-5106285" [ 741.862696] env[65121]: _type = "Task" [ 741.862696] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.871972] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106285, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.912493] env[65121]: DEBUG nova.compute.manager [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Received event network-changed-309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 741.915221] env[65121]: DEBUG nova.compute.manager [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Refreshing instance network info cache due to event network-changed-309f55e1-46cc-4806-9645-15db4d937a0b. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 741.915221] env[65121]: DEBUG oslo_concurrency.lockutils [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Acquiring lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.915221] env[65121]: DEBUG oslo_concurrency.lockutils [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Acquired lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.915221] env[65121]: DEBUG nova.network.neutron [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Refreshing network info cache for port 309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 742.162042] env[65121]: DEBUG oslo_vmware.api [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106284, 'name': PowerOffVM_Task, 'duration_secs': 0.239818} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.165627] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 742.165846] env[65121]: DEBUG nova.compute.manager [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 742.167275] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66ab646-ea77-4d47-8580-09a70475d093 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.269174] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106282, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.375863] env[65121]: DEBUG oslo_vmware.api [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106285, 'name': PowerOnVM_Task, 'duration_secs': 0.488851} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.379222] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 742.379315] env[65121]: INFO nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Took 8.92 seconds to spawn the instance on the hypervisor. [ 742.379473] env[65121]: DEBUG nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 742.381255] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b95487a-8840-4c78-afda-1fc33eca298f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.419020] env[65121]: WARNING neutronclient.v2_0.client [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.419020] env[65121]: WARNING openstack [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.419020] env[65121]: WARNING openstack [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.535522] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7d2715-22a4-4a05-a0ca-d457a793df26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.548981] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae70b79-3eca-41c7-b5d5-c6d0b2f970b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.582922] env[65121]: WARNING openstack [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.583753] env[65121]: WARNING openstack [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.591838] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e24e81c-abde-4589-922d-47d619618e00 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.606716] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897728ee-ee93-42a8-9cb2-1af2d5a7184b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.620925] env[65121]: DEBUG nova.compute.provider_tree [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.683597] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dabb37f5-4bd8-487c-8454-73c0dfff210a tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.069s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.733401] env[65121]: WARNING neutronclient.v2_0.client [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.734637] env[65121]: WARNING openstack [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.735012] env[65121]: WARNING openstack [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.765212] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106282, 'name': RemoveSnapshot_Task} progress is 12%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.862274] env[65121]: DEBUG nova.network.neutron [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updated VIF entry in instance network info cache for port 309f55e1-46cc-4806-9645-15db4d937a0b. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 742.862869] env[65121]: DEBUG nova.network.neutron [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updating instance_info_cache with network_info: [{"id": "309f55e1-46cc-4806-9645-15db4d937a0b", "address": "fa:16:3e:29:ad:aa", "network": {"id": "08c3e4c7-f714-42eb-af35-74d1d1f92501", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1303322131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "0f0e2921e8704997ba4a1f5cf982989f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309f55e1-46", "ovs_interfaceid": "309f55e1-46cc-4806-9645-15db4d937a0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.894030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "f694fcb6-053b-4649-ac63-7fa98b1373eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.894281] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.902528] env[65121]: INFO nova.compute.manager [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Took 44.62 seconds to build instance. [ 743.124698] env[65121]: DEBUG nova.scheduler.client.report [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 743.265420] env[65121]: DEBUG oslo_vmware.api [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106282, 'name': RemoveSnapshot_Task, 'duration_secs': 1.641843} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.265730] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 743.265921] env[65121]: INFO nova.compute.manager [None req-4c2523a3-cbe0-48db-bda2-b75ece82b9cf tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Took 18.43 seconds to snapshot the instance on the hypervisor. [ 743.366185] env[65121]: DEBUG oslo_concurrency.lockutils [req-1460a1b3-7a3d-44fd-b09b-39ec473d1ea1 req-969bf6a0-be04-4da6-a05c-6c10be73e019 service nova] Releasing lock "refresh_cache-06a99a15-5c8a-4be0-b393-80a104b9ba76" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.404762] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59497764-a442-403f-9962-f5c6d123257f tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.340s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.630596] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.751s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.631122] env[65121]: DEBUG nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 743.633795] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.205s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.635213] env[65121]: INFO nova.compute.claims [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.659567] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a110d855-463f-4db0-af09-d9e94530a438 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.667466] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc383a32-e81e-46c1-9424-da7bbbd9ed01 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Suspending the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 743.669174] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-3a158bdc-b495-400b-b3c1-9015964ba91d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.679297] env[65121]: DEBUG oslo_vmware.api [None req-cc383a32-e81e-46c1-9424-da7bbbd9ed01 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 743.679297] env[65121]: value = "task-5106286" [ 743.679297] env[65121]: _type = "Task" [ 743.679297] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.692055] env[65121]: DEBUG oslo_vmware.api [None req-cc383a32-e81e-46c1-9424-da7bbbd9ed01 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106286, 'name': SuspendVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.908484] env[65121]: DEBUG nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 744.142140] env[65121]: DEBUG nova.compute.utils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 744.143595] env[65121]: DEBUG nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 744.143801] env[65121]: DEBUG nova.network.neutron [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 744.144172] env[65121]: WARNING neutronclient.v2_0.client [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.144934] env[65121]: WARNING neutronclient.v2_0.client [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.145760] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.146819] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.158961] env[65121]: DEBUG nova.objects.instance [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lazy-loading 'flavor' on Instance uuid c2f9fa49-4273-4885-b59d-0635ed739d65 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 744.181195] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.181642] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.181921] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.182270] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.182535] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.191967] env[65121]: INFO nova.compute.manager [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Terminating instance [ 744.200815] env[65121]: DEBUG oslo_vmware.api [None req-cc383a32-e81e-46c1-9424-da7bbbd9ed01 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106286, 'name': SuspendVM_Task} progress is 62%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.224553] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "06a99a15-5c8a-4be0-b393-80a104b9ba76" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.225019] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.225208] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "06a99a15-5c8a-4be0-b393-80a104b9ba76-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.225447] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.225683] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.231149] env[65121]: DEBUG nova.policy [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd57eadfd8e5c4ac2a3dce9129ccff90c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d4597cbb8f54cbfba1946c0d5a5f354', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 744.235931] env[65121]: INFO nova.compute.manager [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Terminating instance [ 744.435774] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.616635] env[65121]: DEBUG nova.network.neutron [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Successfully created port: 356c5d3b-ab25-49df-a05d-8422b8112405 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 744.659047] env[65121]: DEBUG nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 744.669412] env[65121]: DEBUG oslo_concurrency.lockutils [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.669412] env[65121]: DEBUG oslo_concurrency.lockutils [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquired lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.669412] env[65121]: DEBUG nova.network.neutron [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 744.669412] env[65121]: DEBUG nova.objects.instance [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lazy-loading 'info_cache' on Instance uuid c2f9fa49-4273-4885-b59d-0635ed739d65 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 744.698239] env[65121]: DEBUG nova.compute.manager [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 744.698485] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.698900] env[65121]: DEBUG oslo_vmware.api [None req-cc383a32-e81e-46c1-9424-da7bbbd9ed01 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106286, 'name': SuspendVM_Task, 'duration_secs': 0.682614} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.700333] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a8f138-58b5-4777-b654-1e6f72c14cfd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.703078] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc383a32-e81e-46c1-9424-da7bbbd9ed01 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Suspended the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 744.703290] env[65121]: DEBUG nova.compute.manager [None req-cc383a32-e81e-46c1-9424-da7bbbd9ed01 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 744.704320] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371a09bb-5e35-411e-9d96-e792274c7a24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.720711] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 744.723875] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c3f9cd7-8b3e-41fb-bfcc-b910fde65c3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.732370] env[65121]: DEBUG oslo_vmware.api [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 744.732370] env[65121]: value = "task-5106287" [ 744.732370] env[65121]: _type = "Task" [ 744.732370] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.738847] env[65121]: DEBUG nova.compute.manager [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 744.738847] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.743156] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349afad1-cfaa-4474-882d-473c4fd74ba8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.746321] env[65121]: DEBUG oslo_vmware.api [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.758959] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 744.760526] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf6cad28-9618-435e-b334-9b3d9e73ee79 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.769440] env[65121]: DEBUG oslo_vmware.api [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 744.769440] env[65121]: value = "task-5106288" [ 744.769440] env[65121]: _type = "Task" [ 744.769440] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.783337] env[65121]: DEBUG oslo_vmware.api [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.836837] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa29535-85bc-430a-bee9-27d88b049193 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.849415] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc7cafa-9cd8-45d1-8560-8238b41fc827 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.885551] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba11c88-c463-44fa-8001-e846fdaecd1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.895533] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22da459c-07b2-4bc9-91bc-55ef7c70bece {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.911154] env[65121]: DEBUG nova.compute.provider_tree [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.173349] env[65121]: DEBUG nova.objects.base [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 745.242117] env[65121]: DEBUG oslo_vmware.api [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106287, 'name': PowerOffVM_Task, 'duration_secs': 0.207453} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.242842] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 745.242842] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 745.242842] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e52a9e7-e7e4-4c61-8737-757d78b7bd36 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.281624] env[65121]: DEBUG oslo_vmware.api [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106288, 'name': PowerOffVM_Task, 'duration_secs': 0.253446} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.281985] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 745.282178] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 745.282464] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a342537a-f418-425d-b0a9-b6ecfce6df29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.317418] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 745.317634] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 745.317832] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Deleting the datastore file [datastore2] 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 745.318095] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29f2c2f0-4514-457e-b125-931e07a81df3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.327637] env[65121]: DEBUG oslo_vmware.api [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for the task: (returnval){ [ 745.327637] env[65121]: value = "task-5106291" [ 745.327637] env[65121]: _type = "Task" [ 745.327637] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.337915] env[65121]: DEBUG oslo_vmware.api [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.361843] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 745.362063] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 745.362300] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Deleting the datastore file [datastore2] 06a99a15-5c8a-4be0-b393-80a104b9ba76 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 745.362596] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09428b4e-ea44-455a-9fc9-3842acee0dcf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.371646] env[65121]: DEBUG oslo_vmware.api [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for the task: (returnval){ [ 745.371646] env[65121]: value = "task-5106292" [ 745.371646] env[65121]: _type = "Task" [ 745.371646] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.382573] env[65121]: DEBUG oslo_vmware.api [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.417078] env[65121]: DEBUG nova.scheduler.client.report [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 745.670248] env[65121]: DEBUG nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 745.676280] env[65121]: WARNING neutronclient.v2_0.client [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.676931] env[65121]: WARNING openstack [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.677285] env[65121]: WARNING openstack [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.694569] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 745.694852] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 745.694966] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 745.695159] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 745.695297] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 745.695432] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 745.695655] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.695834] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 745.696130] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 745.696216] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 745.696306] env[65121]: DEBUG nova.virt.hardware [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 745.697213] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51637842-e179-4c73-afde-4e606f27643c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.706747] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eacf9ae8-a56f-4904-a272-324d4ea711f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.759733] env[65121]: DEBUG nova.compute.manager [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 745.766947] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a022958-9ee5-4a88-a3e5-0d00c1863a26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.790493] env[65121]: WARNING openstack [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.790908] env[65121]: WARNING openstack [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.839617] env[65121]: DEBUG oslo_vmware.api [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Task: {'id': task-5106291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180849} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.839878] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.840683] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 745.840683] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.840683] env[65121]: INFO nova.compute.manager [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Took 1.14 seconds to destroy the instance on the hypervisor. [ 745.840683] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 745.840932] env[65121]: DEBUG nova.compute.manager [-] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 745.840932] env[65121]: DEBUG nova.network.neutron [-] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 745.841131] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.841806] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.841902] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.883887] env[65121]: DEBUG oslo_vmware.api [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Task: {'id': task-5106292, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160376} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.884158] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.884337] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 745.884541] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.884723] env[65121]: INFO nova.compute.manager [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Took 1.15 seconds to destroy the instance on the hypervisor. [ 745.884957] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 745.885164] env[65121]: DEBUG nova.compute.manager [-] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 745.885258] env[65121]: DEBUG nova.network.neutron [-] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 745.885498] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.886041] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.886296] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.894120] env[65121]: WARNING neutronclient.v2_0.client [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.894749] env[65121]: WARNING openstack [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.895102] env[65121]: WARNING openstack [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.922090] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.288s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 745.922440] env[65121]: DEBUG nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 745.927153] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.928055] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.219s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.929505] env[65121]: INFO nova.compute.claims [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.992377] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.256926] env[65121]: DEBUG nova.network.neutron [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Successfully updated port: 356c5d3b-ab25-49df-a05d-8422b8112405 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 746.279578] env[65121]: INFO nova.compute.manager [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] instance snapshotting [ 746.282300] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc6b86f-ec61-4010-82f6-cf953cc4089d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.304957] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b186b62-a177-41f7-8283-4a3ddf73d9f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.435133] env[65121]: DEBUG nova.compute.utils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 746.440386] env[65121]: DEBUG nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 746.440386] env[65121]: DEBUG nova.network.neutron [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 746.440386] env[65121]: WARNING neutronclient.v2_0.client [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.440557] env[65121]: WARNING neutronclient.v2_0.client [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.442021] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.442021] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.533047] env[65121]: DEBUG nova.network.neutron [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Updating instance_info_cache with network_info: [{"id": "ea99cbd9-05be-40d9-952c-9bee73993946", "address": "fa:16:3e:c1:43:51", "network": {"id": "ee9eeb80-7439-47b0-90fd-0d51c67087f1", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1752594550-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3c0dbd078f6b42d5ad8eeb0a45cac9f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea99cbd9-05", "ovs_interfaceid": "ea99cbd9-05be-40d9-952c-9bee73993946", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.538164] env[65121]: DEBUG nova.policy [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34188668111b4c80829f50b69d9a705d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a481e8a6d4f54aa8a3cf7b2f38357729', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 746.723574] env[65121]: DEBUG nova.compute.manager [req-c326b37a-480c-4895-8d82-b4a49f9ca542 req-7fa99311-b028-402e-bf0c-20bc0eb12836 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Received event network-vif-plugged-356c5d3b-ab25-49df-a05d-8422b8112405 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 746.723799] env[65121]: DEBUG oslo_concurrency.lockutils [req-c326b37a-480c-4895-8d82-b4a49f9ca542 req-7fa99311-b028-402e-bf0c-20bc0eb12836 service nova] Acquiring lock "3f69040e-7df3-4535-a3a7-90f3348ef346-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.724018] env[65121]: DEBUG oslo_concurrency.lockutils [req-c326b37a-480c-4895-8d82-b4a49f9ca542 req-7fa99311-b028-402e-bf0c-20bc0eb12836 service nova] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.724196] env[65121]: DEBUG oslo_concurrency.lockutils [req-c326b37a-480c-4895-8d82-b4a49f9ca542 req-7fa99311-b028-402e-bf0c-20bc0eb12836 service nova] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.724425] env[65121]: DEBUG nova.compute.manager [req-c326b37a-480c-4895-8d82-b4a49f9ca542 req-7fa99311-b028-402e-bf0c-20bc0eb12836 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] No waiting events found dispatching network-vif-plugged-356c5d3b-ab25-49df-a05d-8422b8112405 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 746.724728] env[65121]: WARNING nova.compute.manager [req-c326b37a-480c-4895-8d82-b4a49f9ca542 req-7fa99311-b028-402e-bf0c-20bc0eb12836 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Received unexpected event network-vif-plugged-356c5d3b-ab25-49df-a05d-8422b8112405 for instance with vm_state building and task_state spawning. [ 746.759854] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "refresh_cache-3f69040e-7df3-4535-a3a7-90f3348ef346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.760047] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquired lock "refresh_cache-3f69040e-7df3-4535-a3a7-90f3348ef346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 746.760536] env[65121]: DEBUG nova.network.neutron [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 746.816915] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 746.817241] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-65d5af86-e9dc-43b3-9861-196823cd158f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.829138] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 746.829138] env[65121]: value = "task-5106293" [ 746.829138] env[65121]: _type = "Task" [ 746.829138] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.840548] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106293, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.863168] env[65121]: DEBUG nova.network.neutron [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Successfully created port: dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 746.940820] env[65121]: DEBUG nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 746.972791] env[65121]: DEBUG nova.network.neutron [-] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 747.007287] env[65121]: DEBUG nova.network.neutron [-] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 747.041441] env[65121]: DEBUG oslo_concurrency.lockutils [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Releasing lock "refresh_cache-c2f9fa49-4273-4885-b59d-0635ed739d65" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.270222] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.270334] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.343073] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106293, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.402102] env[65121]: DEBUG nova.compute.manager [req-2dbc4ca8-f057-4d59-af5e-393276775130 req-35983023-a979-4c21-9b8a-6d0e7995fc66 service nova] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Received event network-vif-deleted-ec7592a3-0a2f-43b5-9c55-2e59d85c17cf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 747.404440] env[65121]: DEBUG nova.network.neutron [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 747.446463] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.447948] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.476787] env[65121]: INFO nova.compute.manager [-] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Took 1.59 seconds to deallocate network for instance. [ 747.509855] env[65121]: INFO nova.compute.manager [-] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Took 1.67 seconds to deallocate network for instance. [ 747.543663] env[65121]: WARNING neutronclient.v2_0.client [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 747.543663] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.543663] env[65121]: WARNING openstack [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.646115] env[65121]: DEBUG nova.network.neutron [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Updating instance_info_cache with network_info: [{"id": "356c5d3b-ab25-49df-a05d-8422b8112405", "address": "fa:16:3e:40:94:7e", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap356c5d3b-ab", "ovs_interfaceid": "356c5d3b-ab25-49df-a05d-8422b8112405", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 747.656018] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3051e115-91b0-4e69-a688-c0dcb2ca37b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.664785] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394fec91-300f-4e55-aa0b-d0a667e9f030 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.699744] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263a4c36-9926-4b7e-8301-44dcc93b0fca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.708878] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f69e34-3f86-4a4e-bd9f-dba05e49886c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.724751] env[65121]: DEBUG nova.compute.provider_tree [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.844426] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106293, 'name': CreateSnapshot_Task, 'duration_secs': 0.627753} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.844891] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 747.845760] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203d4e25-6434-4a3d-9077-0ba9a5cbda26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.968571] env[65121]: DEBUG nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 747.989205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.008965] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 748.009337] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 748.009537] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 748.009763] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 748.009948] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 748.010145] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 748.010570] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.010570] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 748.012569] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 748.012569] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 748.012569] env[65121]: DEBUG nova.virt.hardware [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 748.012569] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7d7d8f-c8cc-4458-811f-e1cf0881d897 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.017234] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.021411] env[65121]: DEBUG nova.compute.manager [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 748.025738] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca32d491-e565-45ad-8984-5f8bc59eaca6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.030406] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f710a5-eb19-4d5f-838e-808000715dfb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.058173] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 748.060832] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2d4a4ba-6cff-4082-9622-ff853d586aff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.067908] env[65121]: DEBUG oslo_vmware.api [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 748.067908] env[65121]: value = "task-5106294" [ 748.067908] env[65121]: _type = "Task" [ 748.067908] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.079033] env[65121]: DEBUG oslo_vmware.api [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106294, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.149908] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Releasing lock "refresh_cache-3f69040e-7df3-4535-a3a7-90f3348ef346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.150440] env[65121]: DEBUG nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Instance network_info: |[{"id": "356c5d3b-ab25-49df-a05d-8422b8112405", "address": "fa:16:3e:40:94:7e", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap356c5d3b-ab", "ovs_interfaceid": "356c5d3b-ab25-49df-a05d-8422b8112405", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 748.150943] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:94:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '356c5d3b-ab25-49df-a05d-8422b8112405', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 748.159145] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Creating folder: Project (3d4597cbb8f54cbfba1946c0d5a5f354). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 748.159459] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98d5646c-1230-4f3c-a8bc-28d53d0aa74c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.173080] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Created folder: Project (3d4597cbb8f54cbfba1946c0d5a5f354) in parent group-v993268. [ 748.173612] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Creating folder: Instances. Parent ref: group-v993377. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 748.174009] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-60bab246-8517-48a7-9809-5178ca35b0c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.188023] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Created folder: Instances in parent group-v993377. [ 748.188416] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 748.188676] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 748.188966] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d14a12f5-ba76-4a14-a551-8a496f6124f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.211969] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 748.211969] env[65121]: value = "task-5106297" [ 748.211969] env[65121]: _type = "Task" [ 748.211969] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.221693] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106297, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.228339] env[65121]: DEBUG nova.scheduler.client.report [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 748.365363] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 748.365754] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f8c33ecc-3f56-4a30-bcba-48fd3a62eced {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.376088] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 748.376088] env[65121]: value = "task-5106298" [ 748.376088] env[65121]: _type = "Task" [ 748.376088] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.385158] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106298, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.556488] env[65121]: INFO nova.compute.manager [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] instance snapshotting [ 748.556723] env[65121]: WARNING nova.compute.manager [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 748.560528] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e322c0c4-3e3d-41ad-bca6-b8a556264c07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.595264] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008e3807-f100-4f80-ad0f-77cbc26027a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.605615] env[65121]: DEBUG oslo_vmware.api [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106294, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.617297] env[65121]: DEBUG nova.network.neutron [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Successfully updated port: dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 748.724141] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106297, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.734299] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.806s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.734942] env[65121]: DEBUG nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 748.738118] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.958s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.738446] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.741275] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.771s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.742799] env[65121]: INFO nova.compute.claims [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.769037] env[65121]: DEBUG oslo_concurrency.lockutils [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.769327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.769556] env[65121]: INFO nova.compute.manager [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Rebooting instance [ 748.785987] env[65121]: INFO nova.scheduler.client.report [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Deleted allocations for instance 6a4841b6-6076-4e9c-a6cf-a658957e2931 [ 748.890751] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106298, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.080600] env[65121]: DEBUG oslo_vmware.api [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106294, 'name': PowerOnVM_Task, 'duration_secs': 0.75961} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.080970] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 749.081141] env[65121]: DEBUG nova.compute.manager [None req-267ec6fd-635d-43f3-98a5-d5459e6c7fc7 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 749.082209] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93f9d83-bc1d-436a-91a3-28a224a87d51 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.120740] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 749.121349] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "refresh_cache-50083ee5-9655-4cab-9d50-04a97baac626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.121497] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquired lock "refresh_cache-50083ee5-9655-4cab-9d50-04a97baac626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.121682] env[65121]: DEBUG nova.network.neutron [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 749.122799] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-64692401-cd1a-4844-8abb-df724e09c21f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.128566] env[65121]: DEBUG nova.compute.manager [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Received event network-changed-356c5d3b-ab25-49df-a05d-8422b8112405 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 749.128566] env[65121]: DEBUG nova.compute.manager [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Refreshing instance network info cache due to event network-changed-356c5d3b-ab25-49df-a05d-8422b8112405. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 749.128566] env[65121]: DEBUG oslo_concurrency.lockutils [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Acquiring lock "refresh_cache-3f69040e-7df3-4535-a3a7-90f3348ef346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.128566] env[65121]: DEBUG oslo_concurrency.lockutils [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Acquired lock "refresh_cache-3f69040e-7df3-4535-a3a7-90f3348ef346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.128566] env[65121]: DEBUG nova.network.neutron [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Refreshing network info cache for port 356c5d3b-ab25-49df-a05d-8422b8112405 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 749.137876] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 749.137876] env[65121]: value = "task-5106299" [ 749.137876] env[65121]: _type = "Task" [ 749.137876] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.148627] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106299, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.223818] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106297, 'name': CreateVM_Task, 'duration_secs': 0.612788} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.224015] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 749.224662] env[65121]: WARNING neutronclient.v2_0.client [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.225017] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.225210] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.225565] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 749.225849] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67c84c93-fbd8-4f85-970b-cea9baec0dc2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.232509] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 749.232509] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52aeb4ff-f07d-7fa2-3a64-95cd865ccc06" [ 749.232509] env[65121]: _type = "Task" [ 749.232509] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.245816] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52aeb4ff-f07d-7fa2-3a64-95cd865ccc06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.249559] env[65121]: DEBUG nova.compute.utils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 749.251013] env[65121]: DEBUG nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 749.251172] env[65121]: DEBUG nova.network.neutron [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 749.251486] env[65121]: WARNING neutronclient.v2_0.client [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.251864] env[65121]: WARNING neutronclient.v2_0.client [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.252495] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.252878] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.296017] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b71a90c5-0c6e-434c-8c1b-47f9f96623cb tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "6a4841b6-6076-4e9c-a6cf-a658957e2931" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.987s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.298016] env[65121]: DEBUG oslo_concurrency.lockutils [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.298229] env[65121]: DEBUG oslo_concurrency.lockutils [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.298435] env[65121]: DEBUG nova.network.neutron [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 749.318034] env[65121]: DEBUG nova.policy [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1c51fc43b7f44ce9b023d8cb769e029', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89110618b6ef4ce68b2ca53e7e168139', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 749.389403] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106298, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.561797] env[65121]: DEBUG nova.compute.manager [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Received event network-vif-plugged-dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 749.562068] env[65121]: DEBUG oslo_concurrency.lockutils [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Acquiring lock "50083ee5-9655-4cab-9d50-04a97baac626-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.562192] env[65121]: DEBUG oslo_concurrency.lockutils [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Lock "50083ee5-9655-4cab-9d50-04a97baac626-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 749.562346] env[65121]: DEBUG oslo_concurrency.lockutils [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Lock "50083ee5-9655-4cab-9d50-04a97baac626-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.562492] env[65121]: DEBUG nova.compute.manager [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] No waiting events found dispatching network-vif-plugged-dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 749.562644] env[65121]: WARNING nova.compute.manager [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Received unexpected event network-vif-plugged-dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 for instance with vm_state building and task_state spawning. [ 749.562800] env[65121]: DEBUG nova.compute.manager [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Received event network-changed-dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 749.562931] env[65121]: DEBUG nova.compute.manager [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Refreshing instance network info cache due to event network-changed-dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 749.563107] env[65121]: DEBUG oslo_concurrency.lockutils [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Acquiring lock "refresh_cache-50083ee5-9655-4cab-9d50-04a97baac626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.627265] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.627544] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.636877] env[65121]: WARNING neutronclient.v2_0.client [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.639205] env[65121]: WARNING openstack [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.639205] env[65121]: WARNING openstack [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.658703] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106299, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.663953] env[65121]: DEBUG nova.network.neutron [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Successfully created port: bb4123fb-a2e7-46f0-b5b1-73cd269f125d {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 749.672686] env[65121]: DEBUG nova.network.neutron [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 749.706291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "a43879c7-7378-4700-9581-8daabdafe5b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.706535] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "a43879c7-7378-4700-9581-8daabdafe5b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 749.706778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "a43879c7-7378-4700-9581-8daabdafe5b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.706898] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "a43879c7-7378-4700-9581-8daabdafe5b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 749.707070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "a43879c7-7378-4700-9581-8daabdafe5b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.709275] env[65121]: INFO nova.compute.manager [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Terminating instance [ 749.744658] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52aeb4ff-f07d-7fa2-3a64-95cd865ccc06, 'name': SearchDatastore_Task, 'duration_secs': 0.011298} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.744992] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 749.745234] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 749.745465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.745634] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.745828] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 749.746111] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35e17445-22dc-47ef-8490-3d1facd343e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.756773] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 749.757047] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 749.757915] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6929001-2533-4a86-9087-151b6ed022ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.762255] env[65121]: DEBUG nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 749.775560] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 749.775560] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5261e6a3-4d09-59f0-11a5-d4dbfcead085" [ 749.775560] env[65121]: _type = "Task" [ 749.775560] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.794022] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.794022] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.804941] env[65121]: WARNING neutronclient.v2_0.client [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.806212] env[65121]: WARNING openstack [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.806962] env[65121]: WARNING openstack [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.835028] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5261e6a3-4d09-59f0-11a5-d4dbfcead085, 'name': SearchDatastore_Task, 'duration_secs': 0.02024} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.835028] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a850c69b-c17f-4735-954e-572b0ea36031 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.841944] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 749.841944] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ceaa4c-340e-1356-f2c0-a1c9c7c67de6" [ 749.841944] env[65121]: _type = "Task" [ 749.841944] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.853277] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ceaa4c-340e-1356-f2c0-a1c9c7c67de6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.891694] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106298, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.047683] env[65121]: WARNING openstack [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.048089] env[65121]: WARNING openstack [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.125182] env[65121]: WARNING neutronclient.v2_0.client [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.125932] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.126649] env[65121]: WARNING openstack [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.163116] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106299, 'name': CreateSnapshot_Task, 'duration_secs': 0.726227} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.167214] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 750.168714] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35af7e5f-3e46-46fa-9642-c3a771f630db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.215397] env[65121]: DEBUG nova.compute.manager [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 750.215790] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 750.217339] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276b0dd2-c459-412d-8683-c9e65d792f96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.225312] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 750.228911] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edc8c3eb-dd07-4b60-b59b-e79942ed0b32 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.242870] env[65121]: DEBUG oslo_vmware.api [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 750.242870] env[65121]: value = "task-5106300" [ 750.242870] env[65121]: _type = "Task" [ 750.242870] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.253374] env[65121]: DEBUG oslo_vmware.api [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106300, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.309821] env[65121]: WARNING openstack [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.310233] env[65121]: WARNING openstack [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.348058] env[65121]: WARNING neutronclient.v2_0.client [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.348864] env[65121]: WARNING openstack [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.349488] env[65121]: WARNING openstack [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.365590] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ceaa4c-340e-1356-f2c0-a1c9c7c67de6, 'name': SearchDatastore_Task, 'duration_secs': 0.012947} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.367770] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.367770] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 3f69040e-7df3-4535-a3a7-90f3348ef346/3f69040e-7df3-4535-a3a7-90f3348ef346.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 750.368153] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3f24775-d779-40c0-8d50-e29506e2d021 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.378968] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 750.378968] env[65121]: value = "task-5106301" [ 750.378968] env[65121]: _type = "Task" [ 750.378968] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.400319] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106301, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.400639] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106298, 'name': CloneVM_Task, 'duration_secs': 1.647296} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.405946] env[65121]: DEBUG nova.network.neutron [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Updating instance_info_cache with network_info: [{"id": "dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9", "address": "fa:16:3e:24:9e:b0", "network": {"id": "398126a1-da3b-4d1b-9587-f3251e6c01b0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1687899732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a481e8a6d4f54aa8a3cf7b2f38357729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfa63fc-6b", "ovs_interfaceid": "dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.405946] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Created linked-clone VM from snapshot [ 750.405946] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2937a56-cc1b-400f-8b5c-81ff40c3c603 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.419937] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Uploading image c23cabf5-e61f-4bee-97cd-12e1e1f2a135 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 750.450934] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 750.450934] env[65121]: value = "vm-993380" [ 750.450934] env[65121]: _type = "VirtualMachine" [ 750.450934] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 750.451439] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c33e05c3-492b-4648-b63e-df1ad9aa0a52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.464849] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lease: (returnval){ [ 750.464849] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529d4254-66fd-6223-ccbf-45716c9479bf" [ 750.464849] env[65121]: _type = "HttpNfcLease" [ 750.464849] env[65121]: } obtained for exporting VM: (result){ [ 750.464849] env[65121]: value = "vm-993380" [ 750.464849] env[65121]: _type = "VirtualMachine" [ 750.464849] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 750.464849] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the lease: (returnval){ [ 750.464849] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529d4254-66fd-6223-ccbf-45716c9479bf" [ 750.464849] env[65121]: _type = "HttpNfcLease" [ 750.464849] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 750.474320] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 750.474320] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529d4254-66fd-6223-ccbf-45716c9479bf" [ 750.474320] env[65121]: _type = "HttpNfcLease" [ 750.474320] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 750.487022] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704450cf-1ac8-4116-9c3d-592898a19703 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.496151] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac56ec76-0de6-4f7d-917c-9888055c1e57 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.536812] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0d9e69-31d9-4284-a32b-71f27986f9c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.545344] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a98eb3-9251-4c89-9245-a930fc236364 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.336077] env[65121]: WARNING neutronclient.v2_0.client [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 751.336667] env[65121]: WARNING openstack [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.337013] env[65121]: WARNING openstack [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.349079] env[65121]: DEBUG nova.network.neutron [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Successfully updated port: bb4123fb-a2e7-46f0-b5b1-73cd269f125d {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 751.358085] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 751.359489] env[65121]: DEBUG nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 751.365020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Releasing lock "refresh_cache-50083ee5-9655-4cab-9d50-04a97baac626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.365020] env[65121]: DEBUG nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Instance network_info: |[{"id": "dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9", "address": "fa:16:3e:24:9e:b0", "network": {"id": "398126a1-da3b-4d1b-9587-f3251e6c01b0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1687899732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a481e8a6d4f54aa8a3cf7b2f38357729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfa63fc-6b", "ovs_interfaceid": "dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 751.365020] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ebe6a7ec-8c23-4e82-bb5e-838e204f7224 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.373671] env[65121]: DEBUG oslo_concurrency.lockutils [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Acquired lock "refresh_cache-50083ee5-9655-4cab-9d50-04a97baac626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.373893] env[65121]: DEBUG nova.network.neutron [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Refreshing network info cache for port dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 751.375277] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:9e:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '27138a4c-60c9-45fb-bf37-4c2f765315a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.383975] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Creating folder: Project (a481e8a6d4f54aa8a3cf7b2f38357729). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.389830] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f0e223c5-f80b-4d09-8b91-dbc52d404333 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.402267] env[65121]: DEBUG nova.compute.provider_tree [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.412359] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106301, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579478} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.412579] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 751.412579] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529d4254-66fd-6223-ccbf-45716c9479bf" [ 751.412579] env[65121]: _type = "HttpNfcLease" [ 751.412579] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 751.412782] env[65121]: DEBUG oslo_vmware.api [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106300, 'name': PowerOffVM_Task, 'duration_secs': 0.227686} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.417121] env[65121]: DEBUG nova.network.neutron [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Updated VIF entry in instance network info cache for port 356c5d3b-ab25-49df-a05d-8422b8112405. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 751.417121] env[65121]: DEBUG nova.network.neutron [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Updating instance_info_cache with network_info: [{"id": "356c5d3b-ab25-49df-a05d-8422b8112405", "address": "fa:16:3e:40:94:7e", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap356c5d3b-ab", "ovs_interfaceid": "356c5d3b-ab25-49df-a05d-8422b8112405", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 751.426808] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 751.427151] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 751.427328] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 751.427564] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 751.427750] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 751.427936] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 751.428218] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.428423] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 751.428639] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 751.428845] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 751.429077] env[65121]: DEBUG nova.virt.hardware [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 751.432994] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 3f69040e-7df3-4535-a3a7-90f3348ef346/3f69040e-7df3-4535-a3a7-90f3348ef346.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 751.433328] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.436574] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 751.436574] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529d4254-66fd-6223-ccbf-45716c9479bf" [ 751.436574] env[65121]: _type = "HttpNfcLease" [ 751.436574] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 751.436574] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 751.436574] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 751.436574] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 751.436574] env[65121]: value = "task-5106303" [ 751.436574] env[65121]: _type = "Task" [ 751.436574] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.436574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e34d7f8-a2b0-4b9e-a9ca-0708a0c45f67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.439907] env[65121]: DEBUG oslo_concurrency.lockutils [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] Releasing lock "refresh_cache-3f69040e-7df3-4535-a3a7-90f3348ef346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.440234] env[65121]: DEBUG nova.compute.manager [req-2603f943-40ff-4a0f-83c6-8d365427b2c8 req-aa33e0d1-0b6c-451e-a36c-95d1b435c995 service nova] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Received event network-vif-deleted-309f55e1-46cc-4806-9645-15db4d937a0b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 751.442048] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33cfc080-9874-4f24-8594-453ee0682594 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.445386] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15f8864-062d-4714-ab12-cc85707dfff2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.450067] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9c85ec0-2f90-427d-903c-37428975e34d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.458467] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Created folder: Project (a481e8a6d4f54aa8a3cf7b2f38357729) in parent group-v993268. [ 751.458700] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Creating folder: Instances. Parent ref: group-v993382. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.460856] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2b81f59-eb35-468b-b3f8-18a16a296577 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.472150] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 751.472150] env[65121]: value = "task-5106305" [ 751.472150] env[65121]: _type = "Task" [ 751.472150] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.473049] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bf39b7-18ee-760d-1ae6-3140e6f551ec/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 751.473305] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bf39b7-18ee-760d-1ae6-3140e6f551ec/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 751.481611] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9296938-d3bb-4036-87f7-0866920bb7ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.488099] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106303, 'name': CloneVM_Task} progress is 23%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.492622] env[65121]: DEBUG nova.network.neutron [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 751.500465] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Created folder: Instances in parent group-v993382. [ 751.500743] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 751.571544] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 751.572656] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5682f615-7f6a-4ee4-8af3-3af8e87eed95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.600190] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106305, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076673} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.601454] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.602537] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95266eac-7684-492d-8a06-db295654e71c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.607303] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.607303] env[65121]: value = "task-5106308" [ 751.607303] env[65121]: _type = "Task" [ 751.607303] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.635626] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 3f69040e-7df3-4535-a3a7-90f3348ef346/3f69040e-7df3-4535-a3a7-90f3348ef346.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.637767] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31eb3754-bbb8-40d0-8c22-e580e7913cc3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.654086] env[65121]: DEBUG nova.compute.manager [req-9a4c1182-df32-443b-a0a5-aa266eb0b868 req-8a569ed9-caff-40d1-b63d-5a630b2524ea service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Received event network-vif-plugged-bb4123fb-a2e7-46f0-b5b1-73cd269f125d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 751.655292] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a4c1182-df32-443b-a0a5-aa266eb0b868 req-8a569ed9-caff-40d1-b63d-5a630b2524ea service nova] Acquiring lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.655292] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a4c1182-df32-443b-a0a5-aa266eb0b868 req-8a569ed9-caff-40d1-b63d-5a630b2524ea service nova] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.655292] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a4c1182-df32-443b-a0a5-aa266eb0b868 req-8a569ed9-caff-40d1-b63d-5a630b2524ea service nova] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.655292] env[65121]: DEBUG nova.compute.manager [req-9a4c1182-df32-443b-a0a5-aa266eb0b868 req-8a569ed9-caff-40d1-b63d-5a630b2524ea service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] No waiting events found dispatching network-vif-plugged-bb4123fb-a2e7-46f0-b5b1-73cd269f125d {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 751.655292] env[65121]: WARNING nova.compute.manager [req-9a4c1182-df32-443b-a0a5-aa266eb0b868 req-8a569ed9-caff-40d1-b63d-5a630b2524ea service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Received unexpected event network-vif-plugged-bb4123fb-a2e7-46f0-b5b1-73cd269f125d for instance with vm_state building and task_state spawning. [ 751.661272] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106308, 'name': CreateVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.666769] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 751.666769] env[65121]: value = "task-5106309" [ 751.666769] env[65121]: _type = "Task" [ 751.666769] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.676263] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106309, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.713698] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-53d48429-139b-4070-8b7a-664808639b1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.864824] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "refresh_cache-f84d7aa1-a90b-4c95-9673-fb1a2af7187c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.865025] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "refresh_cache-f84d7aa1-a90b-4c95-9673-fb1a2af7187c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.865205] env[65121]: DEBUG nova.network.neutron [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 751.888728] env[65121]: WARNING neutronclient.v2_0.client [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 751.889429] env[65121]: WARNING openstack [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.889783] env[65121]: WARNING openstack [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.932822] env[65121]: DEBUG nova.scheduler.client.report [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 751.963929] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106303, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.002589] env[65121]: DEBUG oslo_concurrency.lockutils [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.019430] env[65121]: WARNING openstack [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.019871] env[65121]: WARNING openstack [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.120104] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106308, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.131506] env[65121]: WARNING neutronclient.v2_0.client [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.132242] env[65121]: WARNING openstack [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.132782] env[65121]: WARNING openstack [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.178132] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106309, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.200205] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 752.201283] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 752.201283] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Deleting the datastore file [datastore1] a43879c7-7378-4700-9581-8daabdafe5b2 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 752.201283] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e09693de-d432-4b5c-b3a6-7067d072f9f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.217882] env[65121]: DEBUG oslo_vmware.api [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for the task: (returnval){ [ 752.217882] env[65121]: value = "task-5106310" [ 752.217882] env[65121]: _type = "Task" [ 752.217882] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.229049] env[65121]: DEBUG oslo_vmware.api [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106310, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.276816] env[65121]: DEBUG nova.network.neutron [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Updated VIF entry in instance network info cache for port dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 752.278058] env[65121]: DEBUG nova.network.neutron [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Updating instance_info_cache with network_info: [{"id": "dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9", "address": "fa:16:3e:24:9e:b0", "network": {"id": "398126a1-da3b-4d1b-9587-f3251e6c01b0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1687899732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a481e8a6d4f54aa8a3cf7b2f38357729", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "27138a4c-60c9-45fb-bf37-4c2f765315a3", "external-id": "nsx-vlan-transportzone-736", "segmentation_id": 736, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfa63fc-6b", "ovs_interfaceid": "dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 752.368441] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.369028] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.412962] env[65121]: DEBUG nova.network.neutron [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 752.436644] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.437035] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.445187] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.704s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.445695] env[65121]: DEBUG nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 752.449125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.442s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.449125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.451582] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.131s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.453166] env[65121]: INFO nova.compute.claims [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.467341] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106303, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.483649] env[65121]: INFO nova.scheduler.client.report [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleted allocations for instance 7a0d0769-31c5-4ae1-8520-a744f64d39c6 [ 752.508253] env[65121]: DEBUG nova.compute.manager [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 752.509504] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358981b7-ccb8-49c3-8927-1f0995282936 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.543739] env[65121]: WARNING neutronclient.v2_0.client [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.544712] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.545288] env[65121]: WARNING openstack [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.619733] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106308, 'name': CreateVM_Task, 'duration_secs': 0.65011} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.620251] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.621046] env[65121]: WARNING neutronclient.v2_0.client [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.621211] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.621480] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 752.621814] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 752.622239] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28851b7a-5feb-42c6-bbdf-93b16f9b1db3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.629143] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 752.629143] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52663bc9-beb2-f85d-398c-280fd6eee038" [ 752.629143] env[65121]: _type = "Task" [ 752.629143] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.642025] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52663bc9-beb2-f85d-398c-280fd6eee038, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.657221] env[65121]: DEBUG nova.network.neutron [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Updating instance_info_cache with network_info: [{"id": "bb4123fb-a2e7-46f0-b5b1-73cd269f125d", "address": "fa:16:3e:4c:6e:f1", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb4123fb-a2", "ovs_interfaceid": "bb4123fb-a2e7-46f0-b5b1-73cd269f125d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 752.678254] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106309, 'name': ReconfigVM_Task, 'duration_secs': 0.697324} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.678760] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 3f69040e-7df3-4535-a3a7-90f3348ef346/3f69040e-7df3-4535-a3a7-90f3348ef346.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.680348] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af99f1ea-0570-4c25-b6fc-bc782b7586e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.692356] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 752.692356] env[65121]: value = "task-5106311" [ 752.692356] env[65121]: _type = "Task" [ 752.692356] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.703934] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106311, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.730936] env[65121]: DEBUG oslo_vmware.api [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Task: {'id': task-5106310, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300258} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.731292] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.731698] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 752.731931] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.732121] env[65121]: INFO nova.compute.manager [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Took 2.52 seconds to destroy the instance on the hypervisor. [ 752.732597] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 752.732826] env[65121]: DEBUG nova.compute.manager [-] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 752.732921] env[65121]: DEBUG nova.network.neutron [-] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 752.733221] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.733789] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.734068] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.772778] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.782411] env[65121]: DEBUG oslo_concurrency.lockutils [req-247ab9a1-54c8-4262-bb57-7793703e1247 req-cfb29806-ddb4-41a7-aadb-1ee03d0c74fa service nova] Releasing lock "refresh_cache-50083ee5-9655-4cab-9d50-04a97baac626" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.958940] env[65121]: DEBUG nova.compute.utils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 752.964025] env[65121]: DEBUG nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 752.964025] env[65121]: DEBUG nova.network.neutron [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 752.964025] env[65121]: WARNING neutronclient.v2_0.client [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.964025] env[65121]: WARNING neutronclient.v2_0.client [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.964025] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.964025] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.994201] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106303, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.999349] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f03d10ce-7780-490a-8c9c-5f0f2463278d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "7a0d0769-31c5-4ae1-8520-a744f64d39c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.595s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.053906] env[65121]: DEBUG nova.policy [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7249df0c55d144939da34955675dea07', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e282e89aff534fd99e723f22f8845914', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 753.141461] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52663bc9-beb2-f85d-398c-280fd6eee038, 'name': SearchDatastore_Task, 'duration_secs': 0.017869} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.141936] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.142183] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.142413] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.142637] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.142855] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.143143] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fb81a9f-ad15-4ee2-af05-46bec505185f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.156234] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.156330] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 753.158208] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef7e6ed0-afbe-4779-b9a6-222fe2593cb6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.163207] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "refresh_cache-f84d7aa1-a90b-4c95-9673-fb1a2af7187c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.163901] env[65121]: DEBUG nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Instance network_info: |[{"id": "bb4123fb-a2e7-46f0-b5b1-73cd269f125d", "address": "fa:16:3e:4c:6e:f1", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb4123fb-a2", "ovs_interfaceid": "bb4123fb-a2e7-46f0-b5b1-73cd269f125d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 753.168309] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:6e:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb4123fb-a2e7-46f0-b5b1-73cd269f125d', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.172865] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating folder: Project (89110618b6ef4ce68b2ca53e7e168139). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 753.173795] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 753.173795] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c9df53-69dd-8914-6140-5fa748aa2390" [ 753.173795] env[65121]: _type = "Task" [ 753.173795] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.174348] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-347fb028-e76c-4b19-8245-4c6773c81b2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.186979] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c9df53-69dd-8914-6140-5fa748aa2390, 'name': SearchDatastore_Task, 'duration_secs': 0.01807} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.187907] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-795023a4-c83f-4100-a2b8-d3e7ef324a3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.198778] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 753.198778] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a74d2e-defa-eaa9-f056-38a7f14c4a26" [ 753.198778] env[65121]: _type = "Task" [ 753.198778] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.201822] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created folder: Project (89110618b6ef4ce68b2ca53e7e168139) in parent group-v993268. [ 753.202182] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating folder: Instances. Parent ref: group-v993386. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 753.205663] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-330785ff-b6ed-44d7-b743-33faf19db004 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.212589] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106311, 'name': Rename_Task, 'duration_secs': 0.290162} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.213740] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 753.214555] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6956a7b6-b569-4ac5-acb6-0e770cec60a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.219215] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a74d2e-defa-eaa9-f056-38a7f14c4a26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.222391] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created folder: Instances in parent group-v993386. [ 753.222767] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 753.223081] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 753.223329] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19971c7f-32d0-4951-a976-9a1b6620bf07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.242692] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 753.242692] env[65121]: value = "task-5106314" [ 753.242692] env[65121]: _type = "Task" [ 753.242692] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.248844] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.248844] env[65121]: value = "task-5106315" [ 753.248844] env[65121]: _type = "Task" [ 753.248844] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.252915] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106314, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.262740] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106315, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.433634] env[65121]: DEBUG nova.network.neutron [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Successfully created port: 68b68226-9044-4753-b2b7-cbae3b99512e {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 753.472989] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106303, 'name': CloneVM_Task, 'duration_secs': 1.728226} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.473317] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Created linked-clone VM from snapshot [ 753.475670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2441b147-c581-4bd3-87af-19cc68dbd3bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.484276] env[65121]: DEBUG nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 753.487119] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Uploading image a295432d-48a7-4bd3-a67f-aab9f1351225 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 753.536910] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 753.536910] env[65121]: value = "vm-993383" [ 753.536910] env[65121]: _type = "VirtualMachine" [ 753.536910] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 753.536910] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5109fff3-c08d-4617-a4a2-d24c36a52b3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.539596] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e6b2f9-e6bc-48e1-9729-a40515f839b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.545968] env[65121]: DEBUG nova.compute.manager [req-385aa015-7e3f-4d4e-9178-58c065d966d4 req-49a86b18-5db7-4aa5-824e-efe34d8fd165 service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Received event network-vif-deleted-e6d39f1f-d115-4218-a28a-e2cd0d05868d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 753.546635] env[65121]: INFO nova.compute.manager [req-385aa015-7e3f-4d4e-9178-58c065d966d4 req-49a86b18-5db7-4aa5-824e-efe34d8fd165 service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Neutron deleted interface e6d39f1f-d115-4218-a28a-e2cd0d05868d; detaching it from the instance and deleting it from the info cache [ 753.546635] env[65121]: DEBUG nova.network.neutron [req-385aa015-7e3f-4d4e-9178-58c065d966d4 req-49a86b18-5db7-4aa5-824e-efe34d8fd165 service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.549299] env[65121]: DEBUG nova.network.neutron [-] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.561376] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease: (returnval){ [ 753.561376] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5226da68-a75f-aa3c-c2c7-83b97e4474c3" [ 753.561376] env[65121]: _type = "HttpNfcLease" [ 753.561376] env[65121]: } obtained for exporting VM: (result){ [ 753.561376] env[65121]: value = "vm-993383" [ 753.561376] env[65121]: _type = "VirtualMachine" [ 753.561376] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 753.561954] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the lease: (returnval){ [ 753.561954] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5226da68-a75f-aa3c-c2c7-83b97e4474c3" [ 753.561954] env[65121]: _type = "HttpNfcLease" [ 753.561954] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 753.562702] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Doing hard reboot of VM {{(pid=65121) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 753.563716] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-5a119a32-42ba-479d-9aaa-283f47f3206e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.577212] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 753.577212] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5226da68-a75f-aa3c-c2c7-83b97e4474c3" [ 753.577212] env[65121]: _type = "HttpNfcLease" [ 753.577212] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 753.579444] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 753.579444] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5226da68-a75f-aa3c-c2c7-83b97e4474c3" [ 753.579444] env[65121]: _type = "HttpNfcLease" [ 753.579444] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 753.580999] env[65121]: DEBUG oslo_vmware.api [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 753.580999] env[65121]: value = "task-5106317" [ 753.580999] env[65121]: _type = "Task" [ 753.580999] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.584970] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62d68e4-40e4-49a7-b555-ca078d270834 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.603453] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520ad502-343f-ba59-1c39-0bdcb3ae1a01/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 753.603640] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520ad502-343f-ba59-1c39-0bdcb3ae1a01/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 753.720018] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a74d2e-defa-eaa9-f056-38a7f14c4a26, 'name': SearchDatastore_Task, 'duration_secs': 0.027227} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.727737] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.728051] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 50083ee5-9655-4cab-9d50-04a97baac626/50083ee5-9655-4cab-9d50-04a97baac626.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 753.728801] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7cb4071f-cadb-47f0-b14f-6c48d99a0c2d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.731889] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3924d84b-effd-4e1c-8cd5-88ae5146f3f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.741872] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 753.741872] env[65121]: value = "task-5106318" [ 753.741872] env[65121]: _type = "Task" [ 753.741872] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.767076] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106318, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.767844] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106314, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.779323] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106315, 'name': CreateVM_Task, 'duration_secs': 0.527506} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.779586] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 753.780112] env[65121]: WARNING neutronclient.v2_0.client [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.780579] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.780747] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.781077] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 753.787106] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd7467e2-6242-4c5b-9337-df091b7b2753 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.796622] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 753.796622] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52507122-de52-ede6-1850-576f1c7a20c1" [ 753.796622] env[65121]: _type = "Task" [ 753.796622] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.806792] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52507122-de52-ede6-1850-576f1c7a20c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.053039] env[65121]: INFO nova.compute.manager [-] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Took 1.32 seconds to deallocate network for instance. [ 754.054425] env[65121]: DEBUG nova.compute.manager [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Received event network-changed-bb4123fb-a2e7-46f0-b5b1-73cd269f125d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 754.054645] env[65121]: DEBUG nova.compute.manager [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Refreshing instance network info cache due to event network-changed-bb4123fb-a2e7-46f0-b5b1-73cd269f125d. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 754.054944] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Acquiring lock "refresh_cache-f84d7aa1-a90b-4c95-9673-fb1a2af7187c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.055116] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Acquired lock "refresh_cache-f84d7aa1-a90b-4c95-9673-fb1a2af7187c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.055308] env[65121]: DEBUG nova.network.neutron [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Refreshing network info cache for port bb4123fb-a2e7-46f0-b5b1-73cd269f125d {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 754.063320] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-489d8584-6d50-4d7a-bc1b-06f086fbe7f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.079994] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2940ef-7806-4100-8491-203606a5de7e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.112898] env[65121]: DEBUG oslo_vmware.api [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106317, 'name': ResetVM_Task, 'duration_secs': 0.139192} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.114428] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Did hard reboot of VM {{(pid=65121) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 754.114789] env[65121]: DEBUG nova.compute.manager [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 754.115634] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8d73fe-55ba-4c0b-9edb-d6de231a4c93 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.139075] env[65121]: DEBUG nova.compute.manager [req-385aa015-7e3f-4d4e-9178-58c065d966d4 req-49a86b18-5db7-4aa5-824e-efe34d8fd165 service nova] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Detach interface failed, port_id=e6d39f1f-d115-4218-a28a-e2cd0d05868d, reason: Instance a43879c7-7378-4700-9581-8daabdafe5b2 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 754.261324] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106318, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.270376] env[65121]: DEBUG oslo_vmware.api [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106314, 'name': PowerOnVM_Task, 'duration_secs': 0.703058} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.271151] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 754.271788] env[65121]: INFO nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Took 8.60 seconds to spawn the instance on the hypervisor. [ 754.272449] env[65121]: DEBUG nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 754.273431] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5067e2d3-a56e-4986-9299-fa09a1ee6e52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.314413] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52507122-de52-ede6-1850-576f1c7a20c1, 'name': SearchDatastore_Task, 'duration_secs': 0.014104} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.314941] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.316134] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.317485] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.317485] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.317792] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.319325] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91ad368b-7626-4605-9204-62a7fb177dd2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.338516] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.338807] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 754.340202] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b40ebe5b-0fe3-4ce6-9c61-8ff8abe4e1ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.349831] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 754.349831] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522b5d93-f250-3aff-c985-72b4e6ea3a25" [ 754.349831] env[65121]: _type = "Task" [ 754.349831] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.362502] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522b5d93-f250-3aff-c985-72b4e6ea3a25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.381444] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c365fad-8b55-4cf1-8920-69832e782821 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.395815] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080a6b90-8e81-4af1-9f92-e10084ff77ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.440583] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac8a5f1-0ba5-4e67-a5e7-ecb8ca7c5ea0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.450543] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588a8282-0e4f-4aa6-a7b3-9abc0a3f8536 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.470107] env[65121]: DEBUG nova.compute.provider_tree [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.500446] env[65121]: DEBUG nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 754.563097] env[65121]: WARNING neutronclient.v2_0.client [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 754.564096] env[65121]: WARNING openstack [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.564763] env[65121]: WARNING openstack [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.573999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 754.661499] env[65121]: DEBUG oslo_concurrency.lockutils [None req-26e949c4-e5c0-4003-b3ef-ac40c6e8d4d4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.892s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 754.759165] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106318, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.696071} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.759780] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 50083ee5-9655-4cab-9d50-04a97baac626/50083ee5-9655-4cab-9d50-04a97baac626.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 754.762950] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 754.762950] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be875b79-4a3f-4fa9-86ef-057c5b1f09d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.771025] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 754.771025] env[65121]: value = "task-5106319" [ 754.771025] env[65121]: _type = "Task" [ 754.771025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.785358] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106319, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.807424] env[65121]: INFO nova.compute.manager [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Took 47.78 seconds to build instance. [ 754.864217] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522b5d93-f250-3aff-c985-72b4e6ea3a25, 'name': SearchDatastore_Task, 'duration_secs': 0.06273} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.866010] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cbc997c-eba9-46d6-9c75-0bde843e4db3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.873873] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 754.873873] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521a0c88-4c9b-5404-7a0a-11f30d67d2cd" [ 754.873873] env[65121]: _type = "Task" [ 754.873873] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.885393] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521a0c88-4c9b-5404-7a0a-11f30d67d2cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.961195] env[65121]: WARNING openstack [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.961773] env[65121]: WARNING openstack [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.973789] env[65121]: DEBUG nova.scheduler.client.report [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 755.156237] env[65121]: DEBUG nova.network.neutron [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Successfully updated port: 68b68226-9044-4753-b2b7-cbae3b99512e {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 755.283147] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106319, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086233} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.283627] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.285057] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2487306-4026-46c4-b1cb-15298aff2bf3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.294236] env[65121]: WARNING neutronclient.v2_0.client [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.295025] env[65121]: WARNING openstack [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.295408] env[65121]: WARNING openstack [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.320307] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4de53a76-62da-479f-a568-00a5721f2e5f tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.257s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.330410] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 50083ee5-9655-4cab-9d50-04a97baac626/50083ee5-9655-4cab-9d50-04a97baac626.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.332072] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47e9e26c-ec77-4921-9f78-267091dd56d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.356118] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 755.356118] env[65121]: value = "task-5106320" [ 755.356118] env[65121]: _type = "Task" [ 755.356118] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.367725] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106320, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.390021] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521a0c88-4c9b-5404-7a0a-11f30d67d2cd, 'name': SearchDatastore_Task, 'duration_secs': 0.022884} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.390021] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.390021] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] f84d7aa1-a90b-4c95-9673-fb1a2af7187c/f84d7aa1-a90b-4c95-9673-fb1a2af7187c.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 755.390021] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d242a1e-890f-4bf9-b343-9ce0aff6c944 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.399055] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 755.399055] env[65121]: value = "task-5106321" [ 755.399055] env[65121]: _type = "Task" [ 755.399055] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.410314] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.420147] env[65121]: DEBUG nova.network.neutron [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Updated VIF entry in instance network info cache for port bb4123fb-a2e7-46f0-b5b1-73cd269f125d. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 755.420803] env[65121]: DEBUG nova.network.neutron [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Updating instance_info_cache with network_info: [{"id": "bb4123fb-a2e7-46f0-b5b1-73cd269f125d", "address": "fa:16:3e:4c:6e:f1", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb4123fb-a2", "ovs_interfaceid": "bb4123fb-a2e7-46f0-b5b1-73cd269f125d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 755.483235] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.032s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.483818] env[65121]: DEBUG nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 755.486733] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.989s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.486986] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.489769] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.310s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.492031] env[65121]: INFO nova.compute.claims [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.525337] env[65121]: INFO nova.scheduler.client.report [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Deleted allocations for instance 46f89cfb-14aa-483d-aa8a-bc91356272ca [ 755.659669] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "refresh_cache-41af3ee5-0ee6-4957-bad0-5aef51d55bd5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.659945] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquired lock "refresh_cache-41af3ee5-0ee6-4957-bad0-5aef51d55bd5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.660166] env[65121]: DEBUG nova.network.neutron [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 755.850180] env[65121]: DEBUG nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 755.871869] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106320, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.911918] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106321, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.924241] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a7059f3-8fad-453d-96b5-4b0f42650605 req-a479ce1b-27f1-482a-a98a-6561092b43de service nova] Releasing lock "refresh_cache-f84d7aa1-a90b-4c95-9673-fb1a2af7187c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.997715] env[65121]: DEBUG nova.compute.utils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 756.002752] env[65121]: DEBUG nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 756.003280] env[65121]: DEBUG nova.network.neutron [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 756.003750] env[65121]: WARNING neutronclient.v2_0.client [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.004300] env[65121]: WARNING neutronclient.v2_0.client [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.005895] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.006637] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.037676] env[65121]: DEBUG oslo_concurrency.lockutils [None req-42b926b0-3c59-46c6-9509-6f38cb5456e0 tempest-ImagesOneServerTestJSON-1208848406 tempest-ImagesOneServerTestJSON-1208848406-project-member] Lock "46f89cfb-14aa-483d-aa8a-bc91356272ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.175s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.118183] env[65121]: DEBUG nova.policy [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f1ca61e513f4855a105e40ff37a2d75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '982f66705583488cb369f19160cc2ee5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 756.163024] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.163782] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.221786] env[65121]: DEBUG nova.network.neutron [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 756.254817] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.255482] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.333930] env[65121]: WARNING neutronclient.v2_0.client [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.334851] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.335305] env[65121]: WARNING openstack [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.377730] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106320, 'name': ReconfigVM_Task, 'duration_secs': 0.553339} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.378282] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 50083ee5-9655-4cab-9d50-04a97baac626/50083ee5-9655-4cab-9d50-04a97baac626.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.379682] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.379960] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2a88549-8f8a-4563-a1b1-36449a9d4e67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.388797] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 756.388797] env[65121]: value = "task-5106322" [ 756.388797] env[65121]: _type = "Task" [ 756.388797] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.400119] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106322, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.412446] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649678} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.412803] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] f84d7aa1-a90b-4c95-9673-fb1a2af7187c/f84d7aa1-a90b-4c95-9673-fb1a2af7187c.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 756.413062] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.413375] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0f9c647-7a47-49ab-8089-ddad080546b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.423989] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 756.423989] env[65121]: value = "task-5106323" [ 756.423989] env[65121]: _type = "Task" [ 756.423989] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.433868] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106323, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.471280] env[65121]: DEBUG nova.network.neutron [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Successfully created port: 79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 756.504681] env[65121]: DEBUG nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 756.510071] env[65121]: DEBUG nova.network.neutron [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Updating instance_info_cache with network_info: [{"id": "68b68226-9044-4753-b2b7-cbae3b99512e", "address": "fa:16:3e:ec:da:ed", "network": {"id": "3472b220-2046-4d25-9999-172ba2082e4e", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1523331266-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e282e89aff534fd99e723f22f8845914", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68b68226-90", "ovs_interfaceid": "68b68226-9044-4753-b2b7-cbae3b99512e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.631652] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a53663-13dc-448a-a1f1-ea5736fbaa61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.643194] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c569dda-0428-4849-8783-773402f5d5af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.689025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb681bf-8238-4c41-bf91-6f9f2f0f7398 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.696637] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5728d7c9-c9ed-4142-973d-22884c91cbac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.715681] env[65121]: DEBUG nova.compute.provider_tree [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.900826] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106322, 'name': Rename_Task, 'duration_secs': 0.192428} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.901145] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.901429] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b689da6-0e20-4216-8fa8-bd8fb340de3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.909830] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 756.909830] env[65121]: value = "task-5106324" [ 756.909830] env[65121]: _type = "Task" [ 756.909830] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.921108] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.934604] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106323, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098716} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.934932] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.935910] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d489c4-5da4-497f-b6a8-e5a6f433fb1a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.960935] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] f84d7aa1-a90b-4c95-9673-fb1a2af7187c/f84d7aa1-a90b-4c95-9673-fb1a2af7187c.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.961324] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a87353a2-1127-46c9-89a8-7ea62e13eb4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.983944] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 756.983944] env[65121]: value = "task-5106325" [ 756.983944] env[65121]: _type = "Task" [ 756.983944] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.996966] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106325, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.016912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Releasing lock "refresh_cache-41af3ee5-0ee6-4957-bad0-5aef51d55bd5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.017153] env[65121]: DEBUG nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Instance network_info: |[{"id": "68b68226-9044-4753-b2b7-cbae3b99512e", "address": "fa:16:3e:ec:da:ed", "network": {"id": "3472b220-2046-4d25-9999-172ba2082e4e", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1523331266-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e282e89aff534fd99e723f22f8845914", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68b68226-90", "ovs_interfaceid": "68b68226-9044-4753-b2b7-cbae3b99512e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 757.220040] env[65121]: DEBUG nova.scheduler.client.report [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 757.425893] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.496824] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106325, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.519847] env[65121]: DEBUG nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 757.725756] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.726450] env[65121]: DEBUG nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 757.729446] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.105s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.732075] env[65121]: INFO nova.compute.claims [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.922763] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.997488] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106325, 'name': ReconfigVM_Task, 'duration_secs': 0.569217} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.997762] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Reconfigured VM instance instance-00000028 to attach disk [datastore2] f84d7aa1-a90b-4c95-9673-fb1a2af7187c/f84d7aa1-a90b-4c95-9673-fb1a2af7187c.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.998948] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b8ea085-9596-43dc-aac5-60ad9e44e064 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.007316] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 758.007316] env[65121]: value = "task-5106326" [ 758.007316] env[65121]: _type = "Task" [ 758.007316] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.020789] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106326, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.065370] env[65121]: DEBUG nova.network.neutron [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Successfully updated port: 79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 758.240379] env[65121]: DEBUG nova.compute.utils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 758.242054] env[65121]: DEBUG nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 758.242288] env[65121]: DEBUG nova.network.neutron [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 758.242724] env[65121]: WARNING neutronclient.v2_0.client [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.243182] env[65121]: WARNING neutronclient.v2_0.client [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.243666] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.244080] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.297564] env[65121]: DEBUG nova.policy [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c468565c58704d9084cfebdaa2897492', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce90017874734825b9fbdce06eb12a00', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 758.422289] env[65121]: DEBUG oslo_vmware.api [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106324, 'name': PowerOnVM_Task, 'duration_secs': 1.505354} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.422475] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 758.422652] env[65121]: INFO nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Took 10.45 seconds to spawn the instance on the hypervisor. [ 758.422843] env[65121]: DEBUG nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 758.423719] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b65982-68dc-467b-8de8-e37b6a63c14d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.519823] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106326, 'name': Rename_Task, 'duration_secs': 0.234607} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.520117] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.520496] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac67e64f-5f50-4e5a-9be9-a3503822459c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.530265] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 758.530265] env[65121]: value = "task-5106327" [ 758.530265] env[65121]: _type = "Task" [ 758.530265] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.540159] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106327, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.568442] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.568727] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.568863] env[65121]: DEBUG nova.network.neutron [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 758.655502] env[65121]: DEBUG nova.network.neutron [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Successfully created port: cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 758.755737] env[65121]: DEBUG nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 758.944282] env[65121]: INFO nova.compute.manager [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Took 51.55 seconds to build instance. [ 759.044397] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106327, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.072507] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.073169] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.119391] env[65121]: DEBUG nova.network.neutron [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 759.145584] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.146348] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.233782] env[65121]: WARNING neutronclient.v2_0.client [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.234653] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.235152] env[65121]: WARNING openstack [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.354137] env[65121]: DEBUG nova.network.neutron [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance_info_cache with network_info: [{"id": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "address": "fa:16:3e:c8:d8:39", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79bfb570-6e", "ovs_interfaceid": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 759.375641] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530e78fb-e23b-4914-870f-4a8233b722f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.386284] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e7d404-d34f-4bcd-9b4e-bd0a14b734b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.426020] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887a423e-bbbc-4f8a-aa56-32e68c6aafce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.433758] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee594eab-f236-4c14-9634-2e04abf109ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.448756] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2c8148e-be79-4af5-97b5-ace3e3914bd3 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "50083ee5-9655-4cab-9d50-04a97baac626" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.385s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.449365] env[65121]: DEBUG nova.compute.provider_tree [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.548030] env[65121]: DEBUG oslo_vmware.api [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106327, 'name': PowerOnVM_Task, 'duration_secs': 0.611406} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.548030] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.548435] env[65121]: INFO nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Took 8.19 seconds to spawn the instance on the hypervisor. [ 759.548435] env[65121]: DEBUG nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 759.549706] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-095a5427-a27a-4edd-9dad-022fcc742074 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.766400] env[65121]: DEBUG nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 759.856731] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.857141] env[65121]: DEBUG nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Instance network_info: |[{"id": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "address": "fa:16:3e:c8:d8:39", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79bfb570-6e", "ovs_interfaceid": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 759.953121] env[65121]: DEBUG nova.scheduler.client.report [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 759.956865] env[65121]: DEBUG nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 760.074248] env[65121]: INFO nova.compute.manager [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Took 52.39 seconds to build instance. [ 760.282117] env[65121]: DEBUG nova.network.neutron [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Successfully updated port: cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 760.461235] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.731s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 760.461657] env[65121]: DEBUG nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 760.466975] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.428s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 760.467220] env[65121]: DEBUG nova.objects.instance [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 760.487184] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 760.576682] env[65121]: DEBUG oslo_concurrency.lockutils [None req-565703fd-8cf1-4195-a074-8ac18076441a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.847s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 760.785864] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.786099] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 760.786256] env[65121]: DEBUG nova.network.neutron [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 760.969025] env[65121]: DEBUG nova.compute.utils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 760.970524] env[65121]: DEBUG nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 760.970725] env[65121]: DEBUG nova.network.neutron [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 760.971141] env[65121]: WARNING neutronclient.v2_0.client [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.971378] env[65121]: WARNING neutronclient.v2_0.client [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.971979] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.972336] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.027302] env[65121]: DEBUG nova.policy [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee00a90ab9a347dc8318c8ae4265bac4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd10d4f413f464f87965d4ac97f69a12f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 761.081329] env[65121]: DEBUG nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 761.223956] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 761.224312] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.224435] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 761.224663] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.224823] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 761.224983] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 761.225256] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.225461] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 761.225691] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 761.225865] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 761.226052] env[65121]: DEBUG nova.virt.hardware [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 761.228833] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfba8c1-71de-4bb0-abe9-6417b8f80466 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.239861] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:22:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='8a66939b-83da-49a6-8c3f-9ca9e786ade0',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1368792546',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 761.240131] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.241064] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 761.241064] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.241064] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 761.241064] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 761.241064] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.241386] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 761.241386] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 761.241386] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 761.241517] env[65121]: DEBUG nova.virt.hardware [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 761.243788] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f663a70b-8af4-4c36-a587-1ef40ff646cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.252378] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f5b976-e026-4940-8c47-49cbc60dd18c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.264296] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 761.265189] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.265189] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 761.265189] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.265189] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 761.265516] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 761.265516] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.265714] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 761.266705] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 761.266705] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 761.266705] env[65121]: DEBUG nova.virt.hardware [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 761.279488] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75ddfdb-1099-41f7-8b1a-7b99274f68f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.283104] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:da:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9ee6f9-33be-4f58-8248-694024ec31d4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '68b68226-9044-4753-b2b7-cbae3b99512e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.292915] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Creating folder: Project (e282e89aff534fd99e723f22f8845914). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.294256] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8958d3-2116-4504-ae68-f3c829091b60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.300510] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.300960] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.309630] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6671f6c-73e0-43e2-b422-5f4c4d73b017 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.313333] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bf39b7-18ee-760d-1ae6-3140e6f551ec/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 761.317163] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb3779ef-3fa1-4fad-b249-8ce602d70a4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.322155] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c037041a-d5fd-439d-80ed-5d7924871dac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.336935] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:d8:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79bfb570-6ed2-4c0f-83e1-1f2cfb088e37', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.344785] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 761.345264] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bf39b7-18ee-760d-1ae6-3140e6f551ec/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 761.346120] env[65121]: ERROR oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bf39b7-18ee-760d-1ae6-3140e6f551ec/disk-0.vmdk due to incomplete transfer. [ 761.348244] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.348535] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f87dafa0-0621-4716-886d-104352d5d491 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.351687] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Created folder: Project (e282e89aff534fd99e723f22f8845914) in parent group-v993268. [ 761.351864] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Creating folder: Instances. Parent ref: group-v993389. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.361015] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5e353e5-1b7b-4cbd-bf9f-6b7062e7bb67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.376783] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a240301c-4331-48ef-aff9-84fbe1250a98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.384449] env[65121]: DEBUG nova.network.neutron [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Successfully created port: 4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 761.388213] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.388213] env[65121]: value = "task-5106329" [ 761.388213] env[65121]: _type = "Task" [ 761.388213] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.390379] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520ad502-343f-ba59-1c39-0bdcb3ae1a01/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 761.392435] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc18b07a-4233-4162-b1a8-23d4bbf28fe0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.395359] env[65121]: DEBUG oslo_vmware.rw_handles [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bf39b7-18ee-760d-1ae6-3140e6f551ec/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 761.395572] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Uploaded image c23cabf5-e61f-4bee-97cd-12e1e1f2a135 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 761.397914] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 761.402370] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8aa61852-b2cc-4aff-81da-14eaa6c6281a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.404479] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Created folder: Instances in parent group-v993389. [ 761.404735] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 761.405350] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.406147] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1981b00-136c-4167-804a-470374ed18b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.428296] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 761.428296] env[65121]: value = "task-5106331" [ 761.428296] env[65121]: _type = "Task" [ 761.428296] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.428525] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106329, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.428769] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520ad502-343f-ba59-1c39-0bdcb3ae1a01/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 761.428915] env[65121]: ERROR oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520ad502-343f-ba59-1c39-0bdcb3ae1a01/disk-0.vmdk due to incomplete transfer. [ 761.429827] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-61cc0743-11b2-4488-b277-a3c951b34498 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.438362] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.438362] env[65121]: value = "task-5106332" [ 761.438362] env[65121]: _type = "Task" [ 761.438362] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.445490] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106331, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.447106] env[65121]: DEBUG oslo_vmware.rw_handles [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520ad502-343f-ba59-1c39-0bdcb3ae1a01/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 761.448530] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Uploaded image a295432d-48a7-4bd3-a67f-aab9f1351225 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 761.449355] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 761.452587] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a3cede4b-8db9-4a30-b8e8-1e839ac85d5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.454272] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106332, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.455292] env[65121]: DEBUG nova.network.neutron [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 761.463532] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 761.463532] env[65121]: value = "task-5106333" [ 761.463532] env[65121]: _type = "Task" [ 761.463532] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.475419] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106333, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.480439] env[65121]: DEBUG nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 761.484818] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df032c22-b376-451f-9673-2f807638442a tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.489029] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.174s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.489029] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.489029] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.162s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.489350] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.491486] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.073s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.491654] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.494106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.997s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.496023] env[65121]: INFO nova.compute.claims [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.528022] env[65121]: INFO nova.scheduler.client.report [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Deleted allocations for instance 5854d21a-d1a8-4043-aec8-b37ff25c40e7 [ 761.556228] env[65121]: INFO nova.scheduler.client.report [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted allocations for instance ab27fa56-f672-4096-a8f4-cce5ff4d5460 [ 761.563967] env[65121]: INFO nova.scheduler.client.report [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Deleted allocations for instance 6164d64c-75c5-469e-ac32-667842b3717a [ 761.613422] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 761.903551] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106329, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.941667] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106331, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.951581] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106332, 'name': CreateVM_Task, 'duration_secs': 0.455413} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.951581] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.951581] env[65121]: WARNING neutronclient.v2_0.client [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 761.951581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.951581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 761.951978] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 761.952045] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a7d0552-609d-47da-8dfa-218e5f56d503 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.958770] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 761.958770] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5250abce-f94b-fb73-7c04-634a2401e75f" [ 761.958770] env[65121]: _type = "Task" [ 761.958770] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.972229] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5250abce-f94b-fb73-7c04-634a2401e75f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.980256] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106333, 'name': Destroy_Task, 'duration_secs': 0.50485} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.980551] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Destroyed the VM [ 761.980809] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 761.981095] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e324fd9d-6c91-4dda-8348-eb4d3ad80387 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.992280] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 761.992280] env[65121]: value = "task-5106334" [ 761.992280] env[65121]: _type = "Task" [ 761.992280] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.004959] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106334, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.023098] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "4b6fe349-335e-4202-ab76-04f6ccc036ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.023389] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.023638] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "4b6fe349-335e-4202-ab76-04f6ccc036ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.023798] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.023972] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.026700] env[65121]: INFO nova.compute.manager [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Terminating instance [ 762.042355] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59711a4c-ffd0-4cd9-afde-59067704a4ef tempest-AttachInterfacesUnderV243Test-2017370622 tempest-AttachInterfacesUnderV243Test-2017370622-project-member] Lock "5854d21a-d1a8-4043-aec8-b37ff25c40e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.756s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.065843] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ef331b1-aadc-4fa9-b2a7-5e627f30ad97 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "ab27fa56-f672-4096-a8f4-cce5ff4d5460" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.212s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.074636] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fced21d8-0927-4c51-96c2-c993e15c3797 tempest-MultipleCreateTestJSON-886329717 tempest-MultipleCreateTestJSON-886329717-project-member] Lock "6164d64c-75c5-469e-ac32-667842b3717a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.252s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.223214] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.223616] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.404974] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106329, 'name': CreateVM_Task, 'duration_secs': 0.617159} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.405204] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 762.405729] env[65121]: WARNING neutronclient.v2_0.client [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.406088] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.441905] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106331, 'name': Destroy_Task, 'duration_secs': 0.598248} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.441905] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Destroyed the VM [ 762.442151] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 762.442435] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8f3ccb4d-b2a1-449d-b06f-91c9e9d875c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.450818] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 762.450818] env[65121]: value = "task-5106335" [ 762.450818] env[65121]: _type = "Task" [ 762.450818] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.463766] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106335, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.475502] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5250abce-f94b-fb73-7c04-634a2401e75f, 'name': SearchDatastore_Task, 'duration_secs': 0.013075} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.476115] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.476411] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.476703] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.477084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.477338] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.477643] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.477961] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 762.478262] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-baa36b35-4d1e-4613-9dc6-d12fe7972372 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.481086] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5656db6-2ee7-4e92-9633-c9cd1a4693a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.488241] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 762.488241] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e726a0-88d9-d12c-679b-c87101726d7d" [ 762.488241] env[65121]: _type = "Task" [ 762.488241] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.497582] env[65121]: DEBUG nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 762.500612] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.500886] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.506895] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-349d9bbb-7146-4985-9b19-32a4631f8b76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.517555] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e726a0-88d9-d12c-679b-c87101726d7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.522664] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 762.522664] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b7b91-95a3-851d-29d6-713d443eb75e" [ 762.522664] env[65121]: _type = "Task" [ 762.522664] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.528774] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106334, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.533068] env[65121]: DEBUG nova.compute.manager [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 762.533359] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.534451] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7407baa-be9d-437a-9e48-0c21555e338f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.542122] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 762.542122] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 762.542475] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 762.542617] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 762.542775] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 762.542918] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 762.543142] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.543332] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 762.543563] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 762.543759] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 762.543946] env[65121]: DEBUG nova.virt.hardware [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 762.544913] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c665ab03-16de-4895-8afe-37f8ef4af306 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.554931] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524b7b91-95a3-851d-29d6-713d443eb75e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.560791] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 762.563541] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7df1dce4-1762-4a7e-95b0-8ea77fdd7702 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.566712] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e88b28-694d-4c8f-b5b1-463aa91b67c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.592941] env[65121]: DEBUG oslo_vmware.api [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 762.592941] env[65121]: value = "task-5106336" [ 762.592941] env[65121]: _type = "Task" [ 762.592941] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.602859] env[65121]: DEBUG oslo_vmware.api [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106336, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.967643] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106335, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.000418] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e726a0-88d9-d12c-679b-c87101726d7d, 'name': SearchDatastore_Task, 'duration_secs': 0.038406} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.003556] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.003797] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.003997] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.014418] env[65121]: DEBUG oslo_vmware.api [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106334, 'name': RemoveSnapshot_Task, 'duration_secs': 0.701406} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.017360] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 763.017595] env[65121]: INFO nova.compute.manager [None req-3b8476d3-c42e-4c48-abe4-0f1ee11ce57c tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Took 14.46 seconds to snapshot the instance on the hypervisor. [ 763.037303] env[65121]: WARNING neutronclient.v2_0.client [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.038146] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.038612] env[65121]: WARNING openstack [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.056283] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524b7b91-95a3-851d-29d6-713d443eb75e, 'name': SearchDatastore_Task, 'duration_secs': 0.022894} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.057619] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff1245f3-7c36-48de-8f42-211211833a11 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.065271] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 763.065271] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52da70d9-eb7f-756c-c76a-09e19bc4c3a7" [ 763.065271] env[65121]: _type = "Task" [ 763.065271] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.075209] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52da70d9-eb7f-756c-c76a-09e19bc4c3a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.109024] env[65121]: DEBUG oslo_vmware.api [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106336, 'name': PowerOffVM_Task, 'duration_secs': 0.301072} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.109024] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 763.109024] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 763.109024] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e7b9f57-9af4-4339-80a3-cb3c0efe0ad6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.115881] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655ab2e1-cbb8-43f3-9596-39e71093371c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.123817] env[65121]: DEBUG nova.network.neutron [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Successfully updated port: 4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 763.128582] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49ac8a8-171e-415a-a4b1-2bdd18d11066 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.170670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f36553-2abf-41c0-bea1-3046f728bedb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.176827] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 763.177052] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 763.177232] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Deleting the datastore file [datastore1] 4b6fe349-335e-4202-ab76-04f6ccc036ee {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 763.177872] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-243d57fc-3e72-40dc-b9ff-91ebe124ea6d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.184542] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cef691-af4e-48cc-b884-b79e0e3e6847 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.190249] env[65121]: DEBUG oslo_vmware.api [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for the task: (returnval){ [ 763.190249] env[65121]: value = "task-5106338" [ 763.190249] env[65121]: _type = "Task" [ 763.190249] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.203028] env[65121]: DEBUG nova.compute.provider_tree [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.211508] env[65121]: DEBUG oslo_vmware.api [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.224189] env[65121]: DEBUG nova.compute.manager [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Received event network-vif-plugged-68b68226-9044-4753-b2b7-cbae3b99512e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 763.224785] env[65121]: DEBUG oslo_concurrency.lockutils [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Acquiring lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.224785] env[65121]: DEBUG oslo_concurrency.lockutils [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.225106] env[65121]: DEBUG oslo_concurrency.lockutils [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.225106] env[65121]: DEBUG nova.compute.manager [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] No waiting events found dispatching network-vif-plugged-68b68226-9044-4753-b2b7-cbae3b99512e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 763.225200] env[65121]: WARNING nova.compute.manager [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Received unexpected event network-vif-plugged-68b68226-9044-4753-b2b7-cbae3b99512e for instance with vm_state building and task_state spawning. [ 763.225329] env[65121]: DEBUG nova.compute.manager [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Received event network-changed-68b68226-9044-4753-b2b7-cbae3b99512e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 763.225630] env[65121]: DEBUG nova.compute.manager [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Refreshing instance network info cache due to event network-changed-68b68226-9044-4753-b2b7-cbae3b99512e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 763.225864] env[65121]: DEBUG oslo_concurrency.lockutils [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Acquiring lock "refresh_cache-41af3ee5-0ee6-4957-bad0-5aef51d55bd5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.226014] env[65121]: DEBUG oslo_concurrency.lockutils [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Acquired lock "refresh_cache-41af3ee5-0ee6-4957-bad0-5aef51d55bd5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.226171] env[65121]: DEBUG nova.network.neutron [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Refreshing network info cache for port 68b68226-9044-4753-b2b7-cbae3b99512e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 763.229760] env[65121]: DEBUG nova.network.neutron [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updating instance_info_cache with network_info: [{"id": "cd096b57-4e55-46f3-b323-cc951f405db2", "address": "fa:16:3e:40:1f:eb", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd096b57-4e", "ovs_interfaceid": "cd096b57-4e55-46f3-b323-cc951f405db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 763.463547] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106335, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.577934] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52da70d9-eb7f-756c-c76a-09e19bc4c3a7, 'name': SearchDatastore_Task, 'duration_secs': 0.012556} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.578213] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.578471] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 41af3ee5-0ee6-4957-bad0-5aef51d55bd5/41af3ee5-0ee6-4957-bad0-5aef51d55bd5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 763.578732] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.578910] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.579155] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbcbaa10-bd10-4a92-882b-1b79d78adf87 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.581279] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3f9cabe-f87c-4f9a-86c7-b3f064117ee8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.590669] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 763.590669] env[65121]: value = "task-5106339" [ 763.590669] env[65121]: _type = "Task" [ 763.590669] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.596251] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.596453] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 763.599061] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3de70385-6f1c-4a17-afe6-d1863c1240cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.605308] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.611676] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 763.611676] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52181a39-ada5-76cd-62e7-142d91200bc6" [ 763.611676] env[65121]: _type = "Task" [ 763.611676] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.626467] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52181a39-ada5-76cd-62e7-142d91200bc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.634337] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "refresh_cache-aab9317b-4ee6-48b3-905b-859a5996f33d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.634546] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquired lock "refresh_cache-aab9317b-4ee6-48b3-905b-859a5996f33d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.635144] env[65121]: DEBUG nova.network.neutron [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 763.701357] env[65121]: DEBUG oslo_vmware.api [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Task: {'id': task-5106338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.375222} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.701494] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.701618] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.702028] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.702028] env[65121]: INFO nova.compute.manager [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Took 1.17 seconds to destroy the instance on the hypervisor. [ 763.702184] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 763.702410] env[65121]: DEBUG nova.compute.manager [-] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 763.702542] env[65121]: DEBUG nova.network.neutron [-] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 763.702767] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.703353] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.703620] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.711439] env[65121]: DEBUG nova.scheduler.client.report [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 763.732256] env[65121]: WARNING neutronclient.v2_0.client [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.733388] env[65121]: WARNING openstack [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.733533] env[65121]: WARNING openstack [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.742235] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Releasing lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.742235] env[65121]: DEBUG nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Instance network_info: |[{"id": "cd096b57-4e55-46f3-b323-cc951f405db2", "address": "fa:16:3e:40:1f:eb", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd096b57-4e", "ovs_interfaceid": "cd096b57-4e55-46f3-b323-cc951f405db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 763.743236] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:1f:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd096b57-4e55-46f3-b323-cc951f405db2', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.753077] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 763.753077] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 763.753077] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-86a77419-d5ab-475c-aafd-5c2a044570bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.775477] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.775477] env[65121]: value = "task-5106340" [ 763.775477] env[65121]: _type = "Task" [ 763.775477] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.785226] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106340, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.817184] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.889558] env[65121]: WARNING openstack [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.890184] env[65121]: WARNING openstack [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.968354] env[65121]: DEBUG oslo_vmware.api [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106335, 'name': RemoveSnapshot_Task, 'duration_secs': 1.223812} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.968354] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 763.968354] env[65121]: INFO nova.compute.manager [None req-ec3bb2af-bb00-4312-b4c7-80508fa597c2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Took 17.69 seconds to snapshot the instance on the hypervisor. [ 763.993613] env[65121]: WARNING neutronclient.v2_0.client [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.994285] env[65121]: WARNING openstack [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.994431] env[65121]: WARNING openstack [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.106977] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106339, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.126023] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52181a39-ada5-76cd-62e7-142d91200bc6, 'name': SearchDatastore_Task, 'duration_secs': 0.013588} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.126591] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-779d6a97-e966-46d1-bded-97b82ddf8d92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.141245] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.141778] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.152082] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 764.152082] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529c5b67-149d-3ad4-6029-cfecb5176955" [ 764.152082] env[65121]: _type = "Task" [ 764.152082] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.168889] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529c5b67-149d-3ad4-6029-cfecb5176955, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.218210] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.724s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.218702] env[65121]: DEBUG nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 764.222292] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.403s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.222522] env[65121]: DEBUG nova.objects.instance [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lazy-loading 'resources' on Instance uuid 1e4a9d58-f848-4ff2-8558-4f23c6082a5f {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 764.266286] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "90cfd88c-286e-4c87-bf27-909873359997" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 764.266523] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "90cfd88c-286e-4c87-bf27-909873359997" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.284583] env[65121]: DEBUG nova.network.neutron [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Updated VIF entry in instance network info cache for port 68b68226-9044-4753-b2b7-cbae3b99512e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 764.285562] env[65121]: DEBUG nova.network.neutron [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Updating instance_info_cache with network_info: [{"id": "68b68226-9044-4753-b2b7-cbae3b99512e", "address": "fa:16:3e:ec:da:ed", "network": {"id": "3472b220-2046-4d25-9999-172ba2082e4e", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1523331266-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e282e89aff534fd99e723f22f8845914", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68b68226-90", "ovs_interfaceid": "68b68226-9044-4753-b2b7-cbae3b99512e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 764.293915] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106340, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.321471] env[65121]: DEBUG nova.network.neutron [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 764.405550] env[65121]: DEBUG nova.compute.manager [req-bf779e41-8100-48d7-ab86-2d188fe17db7 req-2db5ef39-bf42-4609-b217-889412816959 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Received event network-vif-plugged-79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 764.405927] env[65121]: DEBUG oslo_concurrency.lockutils [req-bf779e41-8100-48d7-ab86-2d188fe17db7 req-2db5ef39-bf42-4609-b217-889412816959 service nova] Acquiring lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 764.406098] env[65121]: DEBUG oslo_concurrency.lockutils [req-bf779e41-8100-48d7-ab86-2d188fe17db7 req-2db5ef39-bf42-4609-b217-889412816959 service nova] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.406234] env[65121]: DEBUG oslo_concurrency.lockutils [req-bf779e41-8100-48d7-ab86-2d188fe17db7 req-2db5ef39-bf42-4609-b217-889412816959 service nova] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.406410] env[65121]: DEBUG nova.compute.manager [req-bf779e41-8100-48d7-ab86-2d188fe17db7 req-2db5ef39-bf42-4609-b217-889412816959 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] No waiting events found dispatching network-vif-plugged-79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 764.406545] env[65121]: WARNING nova.compute.manager [req-bf779e41-8100-48d7-ab86-2d188fe17db7 req-2db5ef39-bf42-4609-b217-889412816959 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Received unexpected event network-vif-plugged-79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 for instance with vm_state building and task_state spawning. [ 764.467312] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.467783] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.608985] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106339, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.661176] env[65121]: WARNING neutronclient.v2_0.client [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.661826] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.662210] env[65121]: WARNING openstack [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.677635] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529c5b67-149d-3ad4-6029-cfecb5176955, 'name': SearchDatastore_Task, 'duration_secs': 0.06161} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.677829] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.677951] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 764.678294] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a8d3820-d74f-4e97-b0ca-28cd8ed4d057 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.691571] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 764.691571] env[65121]: value = "task-5106341" [ 764.691571] env[65121]: _type = "Task" [ 764.691571] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.702162] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.727366] env[65121]: DEBUG nova.compute.utils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 764.731948] env[65121]: DEBUG nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 764.732817] env[65121]: DEBUG nova.network.neutron [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 764.732817] env[65121]: WARNING neutronclient.v2_0.client [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.732817] env[65121]: WARNING neutronclient.v2_0.client [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.733420] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.733743] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.787818] env[65121]: DEBUG nova.network.neutron [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Updating instance_info_cache with network_info: [{"id": "4dd0a4a1-30fb-4c71-b7d8-e51204984bd4", "address": "fa:16:3e:e4:4c:30", "network": {"id": "783b529c-63f7-4185-8bb5-cf34ac4d029c", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-509663278-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d10d4f413f464f87965d4ac97f69a12f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dd0a4a1-30", "ovs_interfaceid": "4dd0a4a1-30fb-4c71-b7d8-e51204984bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 764.793373] env[65121]: DEBUG oslo_concurrency.lockutils [req-88563c7d-49c4-42cd-95de-9fb6c69716ba req-44490485-8fdd-4562-b23c-7d73db3215f8 service nova] Releasing lock "refresh_cache-41af3ee5-0ee6-4957-bad0-5aef51d55bd5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.801402] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106340, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.823232] env[65121]: DEBUG nova.policy [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26beb81d28f44cc8932ea8e87a4cd2c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0630960dcbf44781be05184565d81932', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 765.019240] env[65121]: DEBUG nova.network.neutron [-] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 765.105022] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106339, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.202741] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.242865] env[65121]: DEBUG nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 765.295469] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Releasing lock "refresh_cache-aab9317b-4ee6-48b3-905b-859a5996f33d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.295864] env[65121]: DEBUG nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Instance network_info: |[{"id": "4dd0a4a1-30fb-4c71-b7d8-e51204984bd4", "address": "fa:16:3e:e4:4c:30", "network": {"id": "783b529c-63f7-4185-8bb5-cf34ac4d029c", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-509663278-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d10d4f413f464f87965d4ac97f69a12f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dd0a4a1-30", "ovs_interfaceid": "4dd0a4a1-30fb-4c71-b7d8-e51204984bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 765.299601] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106340, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.303482] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:4c:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe99da4f-5630-4afd-918b-b327193d8489', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4dd0a4a1-30fb-4c71-b7d8-e51204984bd4', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.313258] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Creating folder: Project (d10d4f413f464f87965d4ac97f69a12f). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.317416] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ff4654c-da78-4c57-bcf8-66fec670cd22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.331499] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Created folder: Project (d10d4f413f464f87965d4ac97f69a12f) in parent group-v993268. [ 765.331835] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Creating folder: Instances. Parent ref: group-v993394. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.335149] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6915480-086a-4741-a24e-f45360183b9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.353144] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Created folder: Instances in parent group-v993394. [ 765.353432] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 765.353638] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 765.353864] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae094e26-85df-4849-aff8-a2ce406cbd35 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.379296] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.379296] env[65121]: value = "task-5106344" [ 765.379296] env[65121]: _type = "Task" [ 765.379296] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.395111] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106344, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.423088] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b38830-5700-4521-9f45-1000190b332c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.431940] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe71d35-26d8-4b5c-8132-4b78decffd89 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.466757] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff84736-3833-4ef2-a50d-b8d003705235 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.476039] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d8256b-b385-49ef-bfc6-b96518c7de80 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.492279] env[65121]: DEBUG nova.compute.provider_tree [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.523375] env[65121]: INFO nova.compute.manager [-] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Took 1.82 seconds to deallocate network for instance. [ 765.606596] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106339, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.786969} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.607953] env[65121]: DEBUG nova.network.neutron [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Successfully created port: d00d6f6e-2d62-4ae9-8df3-0a5f04428080 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 765.610167] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 41af3ee5-0ee6-4957-bad0-5aef51d55bd5/41af3ee5-0ee6-4957-bad0-5aef51d55bd5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 765.610427] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.610961] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4070eb5-e275-4a16-aac8-d1e1734f01b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.619312] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 765.619312] env[65121]: value = "task-5106345" [ 765.619312] env[65121]: _type = "Task" [ 765.619312] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.628182] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106345, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.704171] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.794429] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106340, 'name': CreateVM_Task, 'duration_secs': 1.571015} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.795328] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 765.795860] env[65121]: WARNING neutronclient.v2_0.client [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.796227] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.796373] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.796828] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 765.797229] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82757302-3c99-4ed4-9f96-473e3a5b5beb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.802394] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 765.802394] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e57db0-8911-b62a-d5fe-bf77990fc808" [ 765.802394] env[65121]: _type = "Task" [ 765.802394] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.815416] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e57db0-8911-b62a-d5fe-bf77990fc808, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.890212] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106344, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.996529] env[65121]: DEBUG nova.scheduler.client.report [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 766.031325] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.130463] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106345, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.205428] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106341, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.503314} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.205729] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 766.206608] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.206608] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-453659cf-ac1e-4c31-8a45-f34c367d644a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.215265] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 766.215265] env[65121]: value = "task-5106346" [ 766.215265] env[65121]: _type = "Task" [ 766.215265] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.225347] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106346, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.256999] env[65121]: DEBUG nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 766.287913] env[65121]: DEBUG nova.virt.hardware [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 766.289030] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e29a64-5516-403e-9a03-8a8316140fae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.298204] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28e2115-3e2d-4ee2-ab3e-8d409c4129bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.323026] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e57db0-8911-b62a-d5fe-bf77990fc808, 'name': SearchDatastore_Task, 'duration_secs': 0.022672} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.323362] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.323697] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.324044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.324044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.324167] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.324468] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b345bdb-f02b-464d-97bb-fa685420f815 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.334285] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.334504] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 766.335333] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a710b3ea-6dda-45ca-a9f8-55f3f69aa748 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.341726] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 766.341726] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ecfb87-b860-a133-954d-9d82a1cbf446" [ 766.341726] env[65121]: _type = "Task" [ 766.341726] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.352454] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ecfb87-b860-a133-954d-9d82a1cbf446, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.392998] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106344, 'name': CreateVM_Task, 'duration_secs': 0.935872} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.393221] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 766.393756] env[65121]: WARNING neutronclient.v2_0.client [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.394153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.394460] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.394678] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 766.394919] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa866cc1-7466-4047-b137-ee29061e4332 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.400617] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 766.400617] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52556d4b-64dd-9d68-9b8b-b2ec6bce62c5" [ 766.400617] env[65121]: _type = "Task" [ 766.400617] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.410792] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52556d4b-64dd-9d68-9b8b-b2ec6bce62c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.503667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.281s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.506974] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.210s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.507345] env[65121]: DEBUG nova.objects.instance [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lazy-loading 'resources' on Instance uuid aa9b6708-c53c-4117-9b75-9d506f393395 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 766.531483] env[65121]: INFO nova.scheduler.client.report [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Deleted allocations for instance 1e4a9d58-f848-4ff2-8558-4f23c6082a5f [ 766.633699] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106345, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.790901} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.633993] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.634914] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2287d83-b3c7-4501-a195-0f6ce9d76176 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.658915] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 41af3ee5-0ee6-4957-bad0-5aef51d55bd5/41af3ee5-0ee6-4957-bad0-5aef51d55bd5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.659639] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3aad7f1b-a003-4eeb-b49d-1a57838eb75a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.682216] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 766.682216] env[65121]: value = "task-5106347" [ 766.682216] env[65121]: _type = "Task" [ 766.682216] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.695090] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106347, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.729253] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106346, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076109} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.729253] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.729253] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a326b4f-eb98-4e53-af10-9ff0008560c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.753670] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.753991] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b545ed86-629c-4e8c-a1a2-34f8df06d755 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.776207] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 766.776207] env[65121]: value = "task-5106348" [ 766.776207] env[65121]: _type = "Task" [ 766.776207] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.785864] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106348, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.853965] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ecfb87-b860-a133-954d-9d82a1cbf446, 'name': SearchDatastore_Task, 'duration_secs': 0.012087} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.855386] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d91e1278-7981-42d6-b785-7874fdf8d0f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.862523] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 766.862523] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e4eb27-f335-c073-97f3-4962a5e618e5" [ 766.862523] env[65121]: _type = "Task" [ 766.862523] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.872370] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e4eb27-f335-c073-97f3-4962a5e618e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.912224] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52556d4b-64dd-9d68-9b8b-b2ec6bce62c5, 'name': SearchDatastore_Task, 'duration_secs': 0.024789} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.912528] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.912761] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.912963] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.043576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-57b979be-e4b3-4849-84eb-2099fb4456e4 tempest-ServersListShow298Test-2070679785 tempest-ServersListShow298Test-2070679785-project-member] Lock "1e4a9d58-f848-4ff2-8558-4f23c6082a5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.610s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.196220] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106347, 'name': ReconfigVM_Task, 'duration_secs': 0.330585} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.201901] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 41af3ee5-0ee6-4957-bad0-5aef51d55bd5/41af3ee5-0ee6-4957-bad0-5aef51d55bd5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.201901] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91b5c254-f844-40d8-8173-cfdf8ce3f880 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.210655] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 767.210655] env[65121]: value = "task-5106349" [ 767.210655] env[65121]: _type = "Task" [ 767.210655] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.225184] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106349, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.232732] env[65121]: DEBUG nova.network.neutron [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Successfully updated port: d00d6f6e-2d62-4ae9-8df3-0a5f04428080 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 767.292058] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106348, 'name': ReconfigVM_Task, 'duration_secs': 0.339311} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.292058] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.292459] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f7b51ff-6d07-4ac4-b78d-120f345391d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.300447] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 767.300447] env[65121]: value = "task-5106350" [ 767.300447] env[65121]: _type = "Task" [ 767.300447] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.314397] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106350, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.379157] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e4eb27-f335-c073-97f3-4962a5e618e5, 'name': SearchDatastore_Task, 'duration_secs': 0.01879} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.379157] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.379243] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 19b1f3f9-842e-4150-8890-b0b22393c3af/19b1f3f9-842e-4150-8890-b0b22393c3af.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 767.379542] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.381717] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 767.381717] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-111531f7-bd12-4ace-9ac2-5c15856bce64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.382859] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd8d7a54-0fec-48b0-80f7-f3743828685c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.392432] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 767.392432] env[65121]: value = "task-5106351" [ 767.392432] env[65121]: _type = "Task" [ 767.392432] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.407087] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.408603] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 767.408809] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 767.409734] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d9ffebc-e800-4310-8542-0fd6381379c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.420574] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 767.420574] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bcc5ba-b79e-ad51-0beb-532ec705911a" [ 767.420574] env[65121]: _type = "Task" [ 767.420574] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.432331] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bcc5ba-b79e-ad51-0beb-532ec705911a, 'name': SearchDatastore_Task} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.433887] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16ca35ca-d4f7-4647-8354-18079efc1514 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.441013] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 767.441013] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b29bf7-f8f2-9867-4b56-ae57eb35fe51" [ 767.441013] env[65121]: _type = "Task" [ 767.441013] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.453091] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b29bf7-f8f2-9867-4b56-ae57eb35fe51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.648906] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1709e542-d223-496a-8d19-650268dc2f0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.661805] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e53776-4c03-4c4b-8648-b2c42ab0091c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.700514] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c38eb3b-7afe-403e-bb24-850e67866ad8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.709932] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ee4893-2c8c-42f3-9d9b-d855fa150e8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.724908] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106349, 'name': Rename_Task, 'duration_secs': 0.236295} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.733662] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 767.734520] env[65121]: DEBUG nova.compute.provider_tree [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.735411] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e454b372-3dba-4308-9da9-7ccf3ed23d28 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.739941] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-75114f97-fe50-4624-9333-303e411529ea" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.740526] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-75114f97-fe50-4624-9333-303e411529ea" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.740738] env[65121]: DEBUG nova.network.neutron [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 767.748526] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 767.748526] env[65121]: value = "task-5106352" [ 767.748526] env[65121]: _type = "Task" [ 767.748526] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.758179] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106352, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.812742] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106350, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.903248] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.921306] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Received event network-vif-plugged-cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 767.921569] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Acquiring lock "19b1f3f9-842e-4150-8890-b0b22393c3af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.921864] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.922104] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.922211] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] No waiting events found dispatching network-vif-plugged-cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 767.922471] env[65121]: WARNING nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Received unexpected event network-vif-plugged-cd096b57-4e55-46f3-b323-cc951f405db2 for instance with vm_state building and task_state spawning. [ 767.923443] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Received event network-changed-cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 767.923443] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Refreshing instance network info cache due to event network-changed-cd096b57-4e55-46f3-b323-cc951f405db2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 767.923443] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Acquiring lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.923443] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Acquired lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.923443] env[65121]: DEBUG nova.network.neutron [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Refreshing network info cache for port cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 767.952592] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b29bf7-f8f2-9867-4b56-ae57eb35fe51, 'name': SearchDatastore_Task, 'duration_secs': 0.011397} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.952818] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.953156] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] aab9317b-4ee6-48b3-905b-859a5996f33d/aab9317b-4ee6-48b3-905b-859a5996f33d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 767.953443] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5190eb90-3dec-43ec-a1e0-2ca9bc40206d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.961596] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 767.961596] env[65121]: value = "task-5106353" [ 767.961596] env[65121]: _type = "Task" [ 767.961596] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.972992] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.182817] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4b868f78-0e91-4d44-911d-7400a5fcdd39 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "75ef7aa2-4afe-41d6-8e54-be264cfcc350" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.184125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4b868f78-0e91-4d44-911d-7400a5fcdd39 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75ef7aa2-4afe-41d6-8e54-be264cfcc350" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.240889] env[65121]: DEBUG nova.scheduler.client.report [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 768.247448] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.247448] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.264227] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106352, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.313137] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106350, 'name': Rename_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.369266] env[65121]: DEBUG nova.network.neutron [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 768.405444] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.417310] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.417466] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.426570] env[65121]: WARNING neutronclient.v2_0.client [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.427254] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.427597] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.480966] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.745034] env[65121]: DEBUG nova.compute.manager [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Received event network-changed-79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 768.749047] env[65121]: DEBUG nova.compute.manager [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Refreshing instance network info cache due to event network-changed-79bfb570-6ed2-4c0f-83e1-1f2cfb088e37. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 768.749047] env[65121]: DEBUG oslo_concurrency.lockutils [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Acquiring lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.749047] env[65121]: DEBUG oslo_concurrency.lockutils [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Acquired lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.749047] env[65121]: DEBUG nova.network.neutron [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Refreshing network info cache for port 79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 768.755410] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.248s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.762032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.787s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.762032] env[65121]: INFO nova.compute.claims [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.782537] env[65121]: DEBUG oslo_vmware.api [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106352, 'name': PowerOnVM_Task, 'duration_secs': 0.521564} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.782537] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 768.782537] env[65121]: INFO nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Took 14.28 seconds to spawn the instance on the hypervisor. [ 768.782537] env[65121]: DEBUG nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 768.783465] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d1280b-d963-4c83-9a09-9f6233982989 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.794850] env[65121]: INFO nova.scheduler.client.report [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Deleted allocations for instance aa9b6708-c53c-4117-9b75-9d506f393395 [ 768.816825] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106350, 'name': Rename_Task, 'duration_secs': 1.257364} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.816825] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 768.816825] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75a5fe73-0d65-4299-a064-55edf25d1eaa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.827398] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 768.827398] env[65121]: value = "task-5106354" [ 768.827398] env[65121]: _type = "Task" [ 768.827398] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.841092] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106354, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.921555] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106351, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.941047] env[65121]: WARNING neutronclient.v2_0.client [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.941832] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.942372] env[65121]: WARNING openstack [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.976206] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.977740] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.978168] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.203139] env[65121]: DEBUG nova.network.neutron [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Updating instance_info_cache with network_info: [{"id": "d00d6f6e-2d62-4ae9-8df3-0a5f04428080", "address": "fa:16:3e:0f:67:a9", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd00d6f6e-2d", "ovs_interfaceid": "d00d6f6e-2d62-4ae9-8df3-0a5f04428080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.240554] env[65121]: WARNING neutronclient.v2_0.client [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.243041] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.243169] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.255095] env[65121]: WARNING neutronclient.v2_0.client [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.255905] env[65121]: WARNING openstack [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.256398] env[65121]: WARNING openstack [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.319654] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8ea2d6-0759-4041-8f9a-83d62dc76ad5 tempest-FloatingIPsAssociationTestJSON-873144333 tempest-FloatingIPsAssociationTestJSON-873144333-project-member] Lock "aa9b6708-c53c-4117-9b75-9d506f393395" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.822s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.326044] env[65121]: INFO nova.compute.manager [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Took 54.39 seconds to build instance. [ 769.341463] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106354, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.409285] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106351, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.654724} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.409787] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 19b1f3f9-842e-4150-8890-b0b22393c3af/19b1f3f9-842e-4150-8890-b0b22393c3af.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 769.410078] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.410432] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d9c1c95-b1eb-41ce-863c-5259e547885e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.421747] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 769.421747] env[65121]: value = "task-5106355" [ 769.421747] env[65121]: _type = "Task" [ 769.421747] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.441288] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106355, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.477427] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106353, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.627456] env[65121]: DEBUG nova.network.neutron [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updated VIF entry in instance network info cache for port cd096b57-4e55-46f3-b323-cc951f405db2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 769.627906] env[65121]: DEBUG nova.network.neutron [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updating instance_info_cache with network_info: [{"id": "cd096b57-4e55-46f3-b323-cc951f405db2", "address": "fa:16:3e:40:1f:eb", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd096b57-4e", "ovs_interfaceid": "cd096b57-4e55-46f3-b323-cc951f405db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 769.690352] env[65121]: WARNING openstack [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.690352] env[65121]: WARNING openstack [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.704361] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-75114f97-fe50-4624-9333-303e411529ea" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.704747] env[65121]: DEBUG nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Instance network_info: |[{"id": "d00d6f6e-2d62-4ae9-8df3-0a5f04428080", "address": "fa:16:3e:0f:67:a9", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd00d6f6e-2d", "ovs_interfaceid": "d00d6f6e-2d62-4ae9-8df3-0a5f04428080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 769.705230] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:67:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd00d6f6e-2d62-4ae9-8df3-0a5f04428080', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.714256] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating folder: Project (0630960dcbf44781be05184565d81932). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.715050] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5be68461-f249-4417-a7fd-6421c3570fd6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.728022] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created folder: Project (0630960dcbf44781be05184565d81932) in parent group-v993268. [ 769.728397] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating folder: Instances. Parent ref: group-v993397. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.728778] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8073dbb9-b560-4fb3-b445-75a674cc1aad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.746848] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created folder: Instances in parent group-v993397. [ 769.746848] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 769.746848] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75114f97-fe50-4624-9333-303e411529ea] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 769.747908] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ac2aa0d-74d1-4a34-b80f-66ff5fdda4c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.776036] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.776036] env[65121]: value = "task-5106358" [ 769.776036] env[65121]: _type = "Task" [ 769.776036] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.789619] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106358, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.834053] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65140ad-ffe8-48b0-abc5-d25760e50e4d tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.074s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.854110] env[65121]: DEBUG oslo_vmware.api [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106354, 'name': PowerOnVM_Task, 'duration_secs': 0.705868} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.854110] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 769.854110] env[65121]: INFO nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Took 12.33 seconds to spawn the instance on the hypervisor. [ 769.854110] env[65121]: DEBUG nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 769.855930] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d1a5db-28b8-4ce5-9286-afc3f90a3dbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.907044] env[65121]: WARNING neutronclient.v2_0.client [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.907561] env[65121]: WARNING openstack [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.907925] env[65121]: WARNING openstack [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.936162] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106355, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117783} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.936162] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 769.936162] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6511f06d-e354-45b2-867e-a9b3368305b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.964707] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 19b1f3f9-842e-4150-8890-b0b22393c3af/19b1f3f9-842e-4150-8890-b0b22393c3af.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.969564] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8daaf3b4-eec6-489f-ab7d-8665130704d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.998466] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106353, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.555144} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.005614] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] aab9317b-4ee6-48b3-905b-859a5996f33d/aab9317b-4ee6-48b3-905b-859a5996f33d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 770.006353] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.006353] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 770.006353] env[65121]: value = "task-5106359" [ 770.006353] env[65121]: _type = "Task" [ 770.006353] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.010262] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69d623a7-3262-4a7d-b573-96add6a83ba8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.027573] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.030100] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 770.030100] env[65121]: value = "task-5106360" [ 770.030100] env[65121]: _type = "Task" [ 770.030100] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.043188] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106360, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.077332] env[65121]: DEBUG nova.network.neutron [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updated VIF entry in instance network info cache for port 79bfb570-6ed2-4c0f-83e1-1f2cfb088e37. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 770.078664] env[65121]: DEBUG nova.network.neutron [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance_info_cache with network_info: [{"id": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "address": "fa:16:3e:c8:d8:39", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79bfb570-6e", "ovs_interfaceid": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.132322] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Releasing lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.132691] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Received event network-vif-plugged-4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 770.132937] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Acquiring lock "aab9317b-4ee6-48b3-905b-859a5996f33d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.133207] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.133388] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.133579] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] No waiting events found dispatching network-vif-plugged-4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 770.133791] env[65121]: WARNING nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Received unexpected event network-vif-plugged-4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 for instance with vm_state building and task_state spawning. [ 770.133907] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Received event network-changed-4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 770.134070] env[65121]: DEBUG nova.compute.manager [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Refreshing instance network info cache due to event network-changed-4dd0a4a1-30fb-4c71-b7d8-e51204984bd4. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 770.134342] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Acquiring lock "refresh_cache-aab9317b-4ee6-48b3-905b-859a5996f33d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.134413] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Acquired lock "refresh_cache-aab9317b-4ee6-48b3-905b-859a5996f33d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.136040] env[65121]: DEBUG nova.network.neutron [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Refreshing network info cache for port 4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 770.286311] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106358, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.340744] env[65121]: DEBUG nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 770.370849] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "3f69040e-7df3-4535-a3a7-90f3348ef346" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.370849] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.371092] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "3f69040e-7df3-4535-a3a7-90f3348ef346-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.371254] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.371433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.377103] env[65121]: INFO nova.compute.manager [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Terminating instance [ 770.388894] env[65121]: INFO nova.compute.manager [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Took 52.10 seconds to build instance. [ 770.495366] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.495732] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.497695] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23e2a20-6339-49a7-8f1a-97c094259a91 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.507546] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda07b87-b5bf-423d-8dc3-513105bf5466 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.545604] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d0211a-4638-4619-aad4-51ec322802d3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.551804] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.557468] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106360, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.561159] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3dc3ba-15b2-43cf-a366-f8846737e616 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.577528] env[65121]: DEBUG nova.compute.provider_tree [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.581791] env[65121]: DEBUG oslo_concurrency.lockutils [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] Releasing lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.582053] env[65121]: DEBUG nova.compute.manager [req-f3569cb2-3750-428e-8f5d-ce658d172871 req-ea934caf-a14e-47a7-a709-c324089af520 service nova] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Received event network-vif-deleted-813d0618-3f84-4020-9e70-55c1bdcf5c22 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 770.637646] env[65121]: WARNING neutronclient.v2_0.client [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.638359] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.638742] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.788853] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106358, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.865908] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.884021] env[65121]: DEBUG nova.compute.manager [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 770.884021] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.884021] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842b23c2-e51b-47f1-9b02-2018abeb24e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.892170] env[65121]: DEBUG oslo_concurrency.lockutils [None req-85ebae89-daf3-439b-9707-bd752e11fe6f tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.753s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.892762] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 770.894135] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1a729a0-be90-4a82-9b39-9482e6d3c1d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.907426] env[65121]: DEBUG oslo_vmware.api [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 770.907426] env[65121]: value = "task-5106361" [ 770.907426] env[65121]: _type = "Task" [ 770.907426] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.917821] env[65121]: DEBUG oslo_vmware.api [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.028480] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.055174] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106360, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.721364} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.055739] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.056665] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f15acb-1988-4f32-9c46-0e6f09371902 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.087422] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] aab9317b-4ee6-48b3-905b-859a5996f33d/aab9317b-4ee6-48b3-905b-859a5996f33d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.092445] env[65121]: DEBUG nova.scheduler.client.report [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 771.103603] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac0aac46-4be7-400e-b292-81d29d08a132 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.124044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.124849] env[65121]: DEBUG nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 771.127754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.694s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.129410] env[65121]: INFO nova.compute.claims [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.142024] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 771.142024] env[65121]: value = "task-5106362" [ 771.142024] env[65121]: _type = "Task" [ 771.142024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.152726] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106362, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.288217] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106358, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.400680] env[65121]: DEBUG nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 771.419353] env[65121]: DEBUG oslo_vmware.api [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106361, 'name': PowerOffVM_Task, 'duration_secs': 0.265205} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.419753] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 771.419829] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 771.420353] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11a48b23-1a6d-4e9c-bf82-4954c77cbbbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.504066] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 771.504385] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 771.504593] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Deleting the datastore file [datastore2] 3f69040e-7df3-4535-a3a7-90f3348ef346 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.504869] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20a8d940-c694-420e-ab2c-213efe92918d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.515820] env[65121]: DEBUG oslo_vmware.api [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for the task: (returnval){ [ 771.515820] env[65121]: value = "task-5106364" [ 771.515820] env[65121]: _type = "Task" [ 771.515820] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.535630] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106359, 'name': ReconfigVM_Task, 'duration_secs': 1.172354} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.535918] env[65121]: DEBUG oslo_vmware.api [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.536212] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 19b1f3f9-842e-4150-8890-b0b22393c3af/19b1f3f9-842e-4150-8890-b0b22393c3af.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 771.537511] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf00d43f-4df1-42d9-8dd7-d3f5d2808ab0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.545264] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 771.545264] env[65121]: value = "task-5106365" [ 771.545264] env[65121]: _type = "Task" [ 771.545264] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.555112] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106365, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.645522] env[65121]: DEBUG nova.compute.utils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 771.651777] env[65121]: DEBUG nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 771.651777] env[65121]: DEBUG nova.network.neutron [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 771.651777] env[65121]: WARNING neutronclient.v2_0.client [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.651777] env[65121]: WARNING neutronclient.v2_0.client [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.652606] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.653666] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.667162] env[65121]: DEBUG nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 771.681170] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106362, 'name': ReconfigVM_Task, 'duration_secs': 0.396353} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.681388] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Reconfigured VM instance instance-0000002c to attach disk [datastore1] aab9317b-4ee6-48b3-905b-859a5996f33d/aab9317b-4ee6-48b3-905b-859a5996f33d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 771.683442] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0df82ca4-3c04-4e85-ad52-222dc680723c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.695506] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 771.695506] env[65121]: value = "task-5106366" [ 771.695506] env[65121]: _type = "Task" [ 771.695506] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.707324] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106366, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.789826] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106358, 'name': CreateVM_Task, 'duration_secs': 1.750489} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.790034] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75114f97-fe50-4624-9333-303e411529ea] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 771.790774] env[65121]: WARNING neutronclient.v2_0.client [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.791216] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.791400] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.791955] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 771.792051] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91a27fdc-4050-4a0d-b5a8-7459e15cef13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.798647] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 771.798647] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c292e4-6abd-9873-b5e9-edb2a806769b" [ 771.798647] env[65121]: _type = "Task" [ 771.798647] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.808132] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c292e4-6abd-9873-b5e9-edb2a806769b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.935984] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.030995] env[65121]: DEBUG oslo_vmware.api [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Task: {'id': task-5106364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159788} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.031271] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.031447] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 772.031629] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.031841] env[65121]: INFO nova.compute.manager [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Took 1.15 seconds to destroy the instance on the hypervisor. [ 772.032389] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 772.032574] env[65121]: DEBUG nova.compute.manager [-] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 772.032860] env[65121]: DEBUG nova.network.neutron [-] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 772.033455] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 772.034312] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.034727] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.055803] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106365, 'name': Rename_Task, 'duration_secs': 0.252615} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.056099] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.056508] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e8b6f5d-124c-492c-b6af-332844f0d22e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.066881] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 772.066881] env[65121]: value = "task-5106367" [ 772.066881] env[65121]: _type = "Task" [ 772.066881] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.080779] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106367, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.094301] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.094301] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.175065] env[65121]: DEBUG nova.policy [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc3c22580f4442dcae4a2b5245ede996', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '083c4f2aa7bf4167b472db6bebb9d8bd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 772.190415] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "50083ee5-9655-4cab-9d50-04a97baac626" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.190697] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "50083ee5-9655-4cab-9d50-04a97baac626" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.191746] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "50083ee5-9655-4cab-9d50-04a97baac626-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.191746] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "50083ee5-9655-4cab-9d50-04a97baac626-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.191746] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "50083ee5-9655-4cab-9d50-04a97baac626-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.195131] env[65121]: INFO nova.compute.manager [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Terminating instance [ 772.214170] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106366, 'name': Rename_Task, 'duration_secs': 0.173438} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.214587] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.214719] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29c24e82-dc12-49e9-99fa-827c3996543e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.224946] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 772.224946] env[65121]: value = "task-5106368" [ 772.224946] env[65121]: _type = "Task" [ 772.224946] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.241441] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106368, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.310823] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c292e4-6abd-9873-b5e9-edb2a806769b, 'name': SearchDatastore_Task, 'duration_secs': 0.010584} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.313955] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.314257] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.314465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.314619] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.314828] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.315358] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2107ddff-2db5-4d0e-8a54-12d4fef56413 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.326418] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.326732] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.330377] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b584901-d589-4350-be56-339268684cd8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.338096] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 772.338096] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5299a248-6939-81ee-cdf9-6011e162ad51" [ 772.338096] env[65121]: _type = "Task" [ 772.338096] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.350812] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5299a248-6939-81ee-cdf9-6011e162ad51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.523677] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 772.585072] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106367, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.685753] env[65121]: DEBUG nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 772.708841] env[65121]: DEBUG nova.compute.manager [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 772.709131] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 772.710356] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579d6b39-ca86-4c0f-8b9e-e1b6009c111a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.727213] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 772.737258] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1116e69-2e70-4aeb-816c-b35a637957b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.754411] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106368, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.758302] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 772.758302] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 772.758302] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 772.758302] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 772.758524] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 772.758566] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 772.758838] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.760015] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 772.760015] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 772.760015] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 772.760015] env[65121]: DEBUG nova.virt.hardware [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 772.762420] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fa247d-91ad-4e6d-a03a-3e17529ef087 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.766093] env[65121]: DEBUG oslo_vmware.api [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 772.766093] env[65121]: value = "task-5106369" [ 772.766093] env[65121]: _type = "Task" [ 772.766093] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.783213] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c46bc6-6d5a-4835-9e6c-7a66002145e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.793496] env[65121]: DEBUG oslo_vmware.api [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.828227] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7998b7f8-5d20-4da6-b536-f5675defc51f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.838480] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0611d4-c11b-46e7-8afc-39417f980e7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.852160] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5299a248-6939-81ee-cdf9-6011e162ad51, 'name': SearchDatastore_Task, 'duration_secs': 0.013029} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.885512] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87d08d2a-cd38-4b62-ac3d-8e9f3e947f38 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.888896] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb8e66f-f966-42c8-8b35-b9acaa8b9a90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.896642] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 772.896642] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bb09f5-0dd5-84ea-34a9-62dddb561c6c" [ 772.896642] env[65121]: _type = "Task" [ 772.896642] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.904916] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af97dbe-18fd-493a-abd0-f7f75eb92299 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.915356] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bb09f5-0dd5-84ea-34a9-62dddb561c6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.925403] env[65121]: DEBUG nova.compute.provider_tree [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.986750] env[65121]: DEBUG nova.network.neutron [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Successfully created port: 00cda2d7-d8b7-4d99-9030-9f175852e4d9 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 773.042661] env[65121]: WARNING neutronclient.v2_0.client [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.043516] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.043903] env[65121]: WARNING openstack [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.078704] env[65121]: DEBUG oslo_vmware.api [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106367, 'name': PowerOnVM_Task, 'duration_secs': 0.699721} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.078704] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.078856] env[65121]: INFO nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Took 13.31 seconds to spawn the instance on the hypervisor. [ 773.079041] env[65121]: DEBUG nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 773.079900] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9070574e-d842-48f7-9c13-30e2bfea55d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.236872] env[65121]: DEBUG oslo_vmware.api [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106368, 'name': PowerOnVM_Task, 'duration_secs': 0.550153} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.237773] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.237773] env[65121]: INFO nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Took 10.74 seconds to spawn the instance on the hypervisor. [ 773.237773] env[65121]: DEBUG nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 773.238439] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77da70cd-46b5-4dbb-bc72-c18d7ee5cbe6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.285502] env[65121]: DEBUG oslo_vmware.api [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106369, 'name': PowerOffVM_Task, 'duration_secs': 0.315973} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.285780] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.286183] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.286256] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf200aa2-4183-44e9-b4f7-fc89617f13eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.373648] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.373648] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.373648] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Deleting the datastore file [datastore2] 50083ee5-9655-4cab-9d50-04a97baac626 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.373648] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb3f2880-5ada-46c4-b47d-c1f9d840cc56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.380568] env[65121]: DEBUG oslo_vmware.api [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for the task: (returnval){ [ 773.380568] env[65121]: value = "task-5106371" [ 773.380568] env[65121]: _type = "Task" [ 773.380568] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.391424] env[65121]: DEBUG oslo_vmware.api [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.415034] env[65121]: DEBUG nova.network.neutron [-] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 773.415034] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bb09f5-0dd5-84ea-34a9-62dddb561c6c, 'name': SearchDatastore_Task, 'duration_secs': 0.021476} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.415034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.415034] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 75114f97-fe50-4624-9333-303e411529ea/75114f97-fe50-4624-9333-303e411529ea.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.415034] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-401c71e5-3c0a-425e-9597-a8015dd333be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.423221] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 773.423221] env[65121]: value = "task-5106372" [ 773.423221] env[65121]: _type = "Task" [ 773.423221] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.430269] env[65121]: DEBUG nova.scheduler.client.report [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 773.437561] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106372, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.602230] env[65121]: INFO nova.compute.manager [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Took 50.45 seconds to build instance. [ 773.760177] env[65121]: INFO nova.compute.manager [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Took 49.16 seconds to build instance. [ 773.765477] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.766205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.767930] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.768433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.768702] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.772207] env[65121]: INFO nova.compute.manager [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Terminating instance [ 773.893102] env[65121]: DEBUG oslo_vmware.api [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Task: {'id': task-5106371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277525} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.893430] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.893783] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 773.893850] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 773.893979] env[65121]: INFO nova.compute.manager [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Took 1.18 seconds to destroy the instance on the hypervisor. [ 773.894247] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 773.894446] env[65121]: DEBUG nova.compute.manager [-] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 773.894541] env[65121]: DEBUG nova.network.neutron [-] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 773.894825] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.895402] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.895659] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.918642] env[65121]: INFO nova.compute.manager [-] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Took 1.89 seconds to deallocate network for instance. [ 773.937551] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106372, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.942701] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.814s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.942912] env[65121]: DEBUG nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 773.946794] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.959s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.947078] env[65121]: DEBUG nova.objects.instance [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lazy-loading 'resources' on Instance uuid 06a99a15-5c8a-4be0-b393-80a104b9ba76 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 774.003124] env[65121]: DEBUG nova.network.neutron [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Updated VIF entry in instance network info cache for port 4dd0a4a1-30fb-4c71-b7d8-e51204984bd4. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 774.003536] env[65121]: DEBUG nova.network.neutron [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Updating instance_info_cache with network_info: [{"id": "4dd0a4a1-30fb-4c71-b7d8-e51204984bd4", "address": "fa:16:3e:e4:4c:30", "network": {"id": "783b529c-63f7-4185-8bb5-cf34ac4d029c", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-509663278-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d10d4f413f464f87965d4ac97f69a12f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dd0a4a1-30", "ovs_interfaceid": "4dd0a4a1-30fb-4c71-b7d8-e51204984bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 774.104473] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b7d7dee8-5a3c-4087-8874-7af573ad20e5 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.283s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.114335] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.263628] env[65121]: DEBUG oslo_concurrency.lockutils [None req-78e21694-2460-4c08-8e9b-14b9acf39603 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.763s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.285019] env[65121]: DEBUG nova.compute.manager [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 774.285019] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.286355] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4b2e32-dd50-4910-bb0f-8fa74803fd12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.297162] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 774.297496] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83d5e78e-46d9-4fcd-a396-1fab25e7196e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.385941] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 774.386301] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 774.386523] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore2] 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 774.386863] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c665516-e5fe-435f-90ea-bf7fdcd5b976 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.404706] env[65121]: DEBUG oslo_vmware.api [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 774.404706] env[65121]: value = "task-5106374" [ 774.404706] env[65121]: _type = "Task" [ 774.404706] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.414396] env[65121]: DEBUG oslo_vmware.api [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106374, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.427705] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 774.436281] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106372, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675523} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.436600] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 75114f97-fe50-4624-9333-303e411529ea/75114f97-fe50-4624-9333-303e411529ea.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.436901] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.437105] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1aefc9ad-9ad7-4099-908f-d282cef63a0d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.445900] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 774.445900] env[65121]: value = "task-5106375" [ 774.445900] env[65121]: _type = "Task" [ 774.445900] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.450834] env[65121]: DEBUG nova.compute.utils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 774.455771] env[65121]: DEBUG nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 774.456587] env[65121]: DEBUG nova.network.neutron [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 774.456587] env[65121]: WARNING neutronclient.v2_0.client [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.456791] env[65121]: WARNING neutronclient.v2_0.client [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.457978] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.457978] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.475760] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106375, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.506243] env[65121]: DEBUG oslo_concurrency.lockutils [req-d09b08c4-18b6-4971-ad6b-18553ce0236e req-d9d57e7f-4121-471c-96fa-619ea52474c2 service nova] Releasing lock "refresh_cache-aab9317b-4ee6-48b3-905b-859a5996f33d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 774.610347] env[65121]: DEBUG nova.compute.manager [None req-4b868f78-0e91-4d44-911d-7400a5fcdd39 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75ef7aa2-4afe-41d6-8e54-be264cfcc350] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 774.666251] env[65121]: DEBUG nova.compute.manager [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Received event network-vif-plugged-d00d6f6e-2d62-4ae9-8df3-0a5f04428080 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 774.666251] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Acquiring lock "75114f97-fe50-4624-9333-303e411529ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 774.666621] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Lock "75114f97-fe50-4624-9333-303e411529ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.667024] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Lock "75114f97-fe50-4624-9333-303e411529ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.667344] env[65121]: DEBUG nova.compute.manager [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] No waiting events found dispatching network-vif-plugged-d00d6f6e-2d62-4ae9-8df3-0a5f04428080 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 774.667653] env[65121]: WARNING nova.compute.manager [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Received unexpected event network-vif-plugged-d00d6f6e-2d62-4ae9-8df3-0a5f04428080 for instance with vm_state building and task_state spawning. [ 774.668074] env[65121]: DEBUG nova.compute.manager [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Received event network-changed-d00d6f6e-2d62-4ae9-8df3-0a5f04428080 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 774.668370] env[65121]: DEBUG nova.compute.manager [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Refreshing instance network info cache due to event network-changed-d00d6f6e-2d62-4ae9-8df3-0a5f04428080. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 774.668679] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Acquiring lock "refresh_cache-75114f97-fe50-4624-9333-303e411529ea" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.668931] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Acquired lock "refresh_cache-75114f97-fe50-4624-9333-303e411529ea" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 774.669708] env[65121]: DEBUG nova.network.neutron [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Refreshing network info cache for port d00d6f6e-2d62-4ae9-8df3-0a5f04428080 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 774.769268] env[65121]: DEBUG nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 774.788429] env[65121]: DEBUG nova.policy [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e04865d199444224ba983603a17ebb6b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f6094b0e36ba4861949628aae919f102', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 774.880076] env[65121]: DEBUG nova.network.neutron [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Successfully updated port: 00cda2d7-d8b7-4d99-9030-9f175852e4d9 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 774.917572] env[65121]: DEBUG oslo_vmware.api [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106374, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.480719} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.918116] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 774.918299] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 774.918467] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.918631] env[65121]: INFO nova.compute.manager [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Took 0.63 seconds to destroy the instance on the hypervisor. [ 774.918906] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 774.919125] env[65121]: DEBUG nova.compute.manager [-] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 774.919220] env[65121]: DEBUG nova.network.neutron [-] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 774.919456] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.919968] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.920236] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.958879] env[65121]: DEBUG nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 774.961570] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106375, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078787} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.966477] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.967670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab59f421-56ef-4f77-9ed8-1e291a30093e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.992431] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] 75114f97-fe50-4624-9333-303e411529ea/75114f97-fe50-4624-9333-303e411529ea.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.995854] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3f5aaae-4a95-4e8c-8dc8-c17e11066d78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.018044] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 775.018044] env[65121]: value = "task-5106376" [ 775.018044] env[65121]: _type = "Task" [ 775.018044] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.030244] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106376, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.097098] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a555a6e2-5967-46f1-801e-92e3db309e03 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.107932] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85501867-fcc9-448f-bc3f-2c1cbe433d56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.142208] env[65121]: DEBUG nova.compute.manager [None req-4b868f78-0e91-4d44-911d-7400a5fcdd39 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75ef7aa2-4afe-41d6-8e54-be264cfcc350] Instance disappeared before build. {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2482}} [ 775.144338] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0fbe7b-868c-484a-929c-405382df6f2d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.150810] env[65121]: DEBUG nova.network.neutron [-] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 775.158970] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd06ee3-a6ce-4b47-b23d-82b9c97b0b15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.172486] env[65121]: WARNING neutronclient.v2_0.client [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.173188] env[65121]: WARNING openstack [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 775.173527] env[65121]: WARNING openstack [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 775.183801] env[65121]: DEBUG nova.compute.provider_tree [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.183801] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.277879] env[65121]: DEBUG nova.network.neutron [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Successfully created port: 7381b4f6-e691-437c-9b26-59e3192fe6ed {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 775.308324] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.385237] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "refresh_cache-b6f6871d-311c-4adb-824e-2907a12f4224" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.385303] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "refresh_cache-b6f6871d-311c-4adb-824e-2907a12f4224" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 775.385430] env[65121]: DEBUG nova.network.neutron [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 775.533397] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.657573] env[65121]: INFO nova.compute.manager [-] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Took 1.76 seconds to deallocate network for instance. [ 775.664810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4b868f78-0e91-4d44-911d-7400a5fcdd39 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75ef7aa2-4afe-41d6-8e54-be264cfcc350" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 7.482s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.688398] env[65121]: DEBUG nova.scheduler.client.report [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 775.888686] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 775.889140] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 775.957588] env[65121]: WARNING openstack [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 775.958021] env[65121]: WARNING openstack [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 775.972809] env[65121]: DEBUG nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 776.001343] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 776.001629] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 776.001801] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 776.002035] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 776.002229] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 776.002406] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 776.002715] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.002909] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 776.003129] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 776.003323] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 776.003519] env[65121]: DEBUG nova.virt.hardware [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 776.004613] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bcbbff-2b48-4683-a343-98e4ff1948fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.014300] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2216e1e-7110-40f2-9275-ca9e913aaae9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.052047] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106376, 'name': ReconfigVM_Task, 'duration_secs': 0.663559} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.052047] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Reconfigured VM instance instance-0000002d to attach disk [datastore1] 75114f97-fe50-4624-9333-303e411529ea/75114f97-fe50-4624-9333-303e411529ea.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.052047] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-790d85d8-c40b-4ce9-89c2-c2107e2763e9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.062035] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 776.062035] env[65121]: value = "task-5106377" [ 776.062035] env[65121]: _type = "Task" [ 776.062035] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.065383] env[65121]: DEBUG nova.network.neutron [-] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 776.073012] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106377, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.167393] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.182364] env[65121]: DEBUG nova.network.neutron [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 776.195118] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.248s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.197928] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.181s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.198156] env[65121]: DEBUG nova.objects.instance [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lazy-loading 'resources' on Instance uuid 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 776.226565] env[65121]: INFO nova.scheduler.client.report [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Deleted allocations for instance 06a99a15-5c8a-4be0-b393-80a104b9ba76 [ 776.549925] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.550340] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.567889] env[65121]: INFO nova.compute.manager [-] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Took 1.65 seconds to deallocate network for instance. [ 776.573352] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106377, 'name': Rename_Task, 'duration_secs': 0.190656} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.576196] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.576791] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b75e111-2ee4-41e5-a86d-b7907a18b852 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.585046] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 776.585046] env[65121]: value = "task-5106378" [ 776.585046] env[65121]: _type = "Task" [ 776.585046] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.594550] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106378, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.739841] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16ae0031-6e81-4ee4-8391-e1ba1ed50969 tempest-FloatingIPsAssociationNegativeTestJSON-276968950 tempest-FloatingIPsAssociationNegativeTestJSON-276968950-project-member] Lock "06a99a15-5c8a-4be0-b393-80a104b9ba76" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.514s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.951354] env[65121]: DEBUG nova.network.neutron [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Successfully updated port: 7381b4f6-e691-437c-9b26-59e3192fe6ed {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 776.989567] env[65121]: WARNING neutronclient.v2_0.client [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.990261] env[65121]: WARNING openstack [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.990724] env[65121]: WARNING openstack [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.080070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.096603] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106378, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.304806] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3455ef6-af05-4e89-8a5b-1bc0a76805de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.314984] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fb0d7f-905a-464f-9b6c-4bcb20f1d8f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.349414] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885d2bc2-4af7-4d19-9da6-64ebe1db6f55 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.358215] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f241d38-2d11-42b1-aca7-8277437c84fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.373519] env[65121]: DEBUG nova.compute.provider_tree [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.414897] env[65121]: WARNING neutronclient.v2_0.client [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 777.415746] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.416308] env[65121]: WARNING openstack [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.457425] env[65121]: DEBUG nova.network.neutron [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Updated VIF entry in instance network info cache for port d00d6f6e-2d62-4ae9-8df3-0a5f04428080. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 777.457425] env[65121]: DEBUG nova.network.neutron [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Updating instance_info_cache with network_info: [{"id": "d00d6f6e-2d62-4ae9-8df3-0a5f04428080", "address": "fa:16:3e:0f:67:a9", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd00d6f6e-2d", "ovs_interfaceid": "d00d6f6e-2d62-4ae9-8df3-0a5f04428080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.463335] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "refresh_cache-00362477-c89e-4f60-98a4-d4928081d55e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.463335] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquired lock "refresh_cache-00362477-c89e-4f60-98a4-d4928081d55e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.463335] env[65121]: DEBUG nova.network.neutron [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 777.600187] env[65121]: DEBUG oslo_vmware.api [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106378, 'name': PowerOnVM_Task, 'duration_secs': 0.530024} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.600867] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.601104] env[65121]: INFO nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Took 11.34 seconds to spawn the instance on the hypervisor. [ 777.601311] env[65121]: DEBUG nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 777.602173] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26acfccd-7cb7-494e-97d4-56ad7d8e682d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.877109] env[65121]: DEBUG nova.scheduler.client.report [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 777.886308] env[65121]: DEBUG nova.network.neutron [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Updating instance_info_cache with network_info: [{"id": "00cda2d7-d8b7-4d99-9030-9f175852e4d9", "address": "fa:16:3e:55:35:54", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00cda2d7-d8", "ovs_interfaceid": "00cda2d7-d8b7-4d99-9030-9f175852e4d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.962624] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f071fcd-5226-41be-9bb9-eac82754fe34 req-9e8820d9-3d2c-48b7-9f58-3c11c7091f1f service nova] Releasing lock "refresh_cache-75114f97-fe50-4624-9333-303e411529ea" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 777.970071] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.970495] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.028556] env[65121]: DEBUG nova.network.neutron [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 778.124824] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.125896] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.138303] env[65121]: INFO nova.compute.manager [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Took 44.66 seconds to build instance. [ 778.311451] env[65121]: WARNING neutronclient.v2_0.client [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.312287] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.312631] env[65121]: WARNING openstack [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.390460] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.192s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.392792] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "refresh_cache-b6f6871d-311c-4adb-824e-2907a12f4224" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 778.393968] env[65121]: DEBUG nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Instance network_info: |[{"id": "00cda2d7-d8b7-4d99-9030-9f175852e4d9", "address": "fa:16:3e:55:35:54", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00cda2d7-d8", "ovs_interfaceid": "00cda2d7-d8b7-4d99-9030-9f175852e4d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 778.394851] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.821s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.398036] env[65121]: DEBUG nova.objects.instance [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lazy-loading 'resources' on Instance uuid a43879c7-7378-4700-9581-8daabdafe5b2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.399663] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:35:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00cda2d7-d8b7-4d99-9030-9f175852e4d9', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.410155] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 778.413123] env[65121]: DEBUG nova.network.neutron [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Updating instance_info_cache with network_info: [{"id": "7381b4f6-e691-437c-9b26-59e3192fe6ed", "address": "fa:16:3e:3d:a9:02", "network": {"id": "e9a0f265-bbfd-40d4-a9b1-63fbae017794", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-58006937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6094b0e36ba4861949628aae919f102", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7836a5b-a91e-4d3f-8e96-afe024f62bb5", "external-id": "nsx-vlan-transportzone-419", "segmentation_id": 419, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7381b4f6-e6", "ovs_interfaceid": "7381b4f6-e691-437c-9b26-59e3192fe6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 778.418917] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.420523] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea5f00dc-c7e0-4988-9d73-c22d1b5ccff9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.438854] env[65121]: INFO nova.scheduler.client.report [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Deleted allocations for instance 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818 [ 778.450015] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.450015] env[65121]: value = "task-5106379" [ 778.450015] env[65121]: _type = "Task" [ 778.450015] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.461416] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106379, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.468599] env[65121]: DEBUG nova.compute.manager [req-d427b44f-76c6-41b4-b457-eb26d366c4ba req-35d88e92-cfae-49bd-ba40-65ea7f6db113 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Received event network-vif-plugged-00cda2d7-d8b7-4d99-9030-9f175852e4d9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 778.468827] env[65121]: DEBUG oslo_concurrency.lockutils [req-d427b44f-76c6-41b4-b457-eb26d366c4ba req-35d88e92-cfae-49bd-ba40-65ea7f6db113 service nova] Acquiring lock "b6f6871d-311c-4adb-824e-2907a12f4224-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.469128] env[65121]: DEBUG oslo_concurrency.lockutils [req-d427b44f-76c6-41b4-b457-eb26d366c4ba req-35d88e92-cfae-49bd-ba40-65ea7f6db113 service nova] Lock "b6f6871d-311c-4adb-824e-2907a12f4224-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.470037] env[65121]: DEBUG oslo_concurrency.lockutils [req-d427b44f-76c6-41b4-b457-eb26d366c4ba req-35d88e92-cfae-49bd-ba40-65ea7f6db113 service nova] Lock "b6f6871d-311c-4adb-824e-2907a12f4224-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.470037] env[65121]: DEBUG nova.compute.manager [req-d427b44f-76c6-41b4-b457-eb26d366c4ba req-35d88e92-cfae-49bd-ba40-65ea7f6db113 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] No waiting events found dispatching network-vif-plugged-00cda2d7-d8b7-4d99-9030-9f175852e4d9 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 778.470037] env[65121]: WARNING nova.compute.manager [req-d427b44f-76c6-41b4-b457-eb26d366c4ba req-35d88e92-cfae-49bd-ba40-65ea7f6db113 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Received unexpected event network-vif-plugged-00cda2d7-d8b7-4d99-9030-9f175852e4d9 for instance with vm_state building and task_state spawning. [ 778.615906] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "75bf0b85-db28-4efa-a517-d7256873b09d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.615906] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75bf0b85-db28-4efa-a517-d7256873b09d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.643081] env[65121]: DEBUG oslo_concurrency.lockutils [None req-28f6d77e-8d7a-4d7a-9566-e98ead54941e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "75114f97-fe50-4624-9333-303e411529ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.285s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.651973] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "3f3bec5b-2834-497c-a454-a152cb992309" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.652767] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "3f3bec5b-2834-497c-a454-a152cb992309" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.918542] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Releasing lock "refresh_cache-00362477-c89e-4f60-98a4-d4928081d55e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 778.918899] env[65121]: DEBUG nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Instance network_info: |[{"id": "7381b4f6-e691-437c-9b26-59e3192fe6ed", "address": "fa:16:3e:3d:a9:02", "network": {"id": "e9a0f265-bbfd-40d4-a9b1-63fbae017794", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-58006937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6094b0e36ba4861949628aae919f102", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7836a5b-a91e-4d3f-8e96-afe024f62bb5", "external-id": "nsx-vlan-transportzone-419", "segmentation_id": 419, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7381b4f6-e6", "ovs_interfaceid": "7381b4f6-e691-437c-9b26-59e3192fe6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 778.919586] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:a9:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7836a5b-a91e-4d3f-8e96-afe024f62bb5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7381b4f6-e691-437c-9b26-59e3192fe6ed', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.928022] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Creating folder: Project (f6094b0e36ba4861949628aae919f102). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.928709] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6f2e28e-793b-431d-abf2-c68a80216a62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.942756] env[65121]: DEBUG nova.compute.manager [req-19a51932-4c58-4d26-b548-71bf7520af59 req-c986ebbb-210a-4961-a12f-d0dac04b7e45 service nova] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Received event network-vif-deleted-356c5d3b-ab25-49df-a05d-8422b8112405 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 778.943056] env[65121]: DEBUG nova.compute.manager [req-19a51932-4c58-4d26-b548-71bf7520af59 req-c986ebbb-210a-4961-a12f-d0dac04b7e45 service nova] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Received event network-vif-deleted-dbfa63fc-6bee-4e96-8e97-3f9ff5bbb2d9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 778.943275] env[65121]: DEBUG nova.compute.manager [req-19a51932-4c58-4d26-b548-71bf7520af59 req-c986ebbb-210a-4961-a12f-d0dac04b7e45 service nova] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Received event network-vif-deleted-b5cdcd85-1d9d-4eb2-8705-20970c1d609e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 778.949566] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Created folder: Project (f6094b0e36ba4861949628aae919f102) in parent group-v993268. [ 778.949793] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Creating folder: Instances. Parent ref: group-v993401. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.950036] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c3ffcbe-0518-4509-bba9-0dd44901c57c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.952130] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ac33beec-4b3d-47de-b0fd-61eb4312d0b5 tempest-VolumesAssistedSnapshotsTest-1179783743 tempest-VolumesAssistedSnapshotsTest-1179783743-project-member] Lock "4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.771s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.967833] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106379, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.974464] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Created folder: Instances in parent group-v993401. [ 778.974769] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 778.975667] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.975844] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d07a0a82-b22b-4dbd-bbdd-dd697e7e8c84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.009209] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.009209] env[65121]: value = "task-5106382" [ 779.009209] env[65121]: _type = "Task" [ 779.009209] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.025359] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106382, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.058034] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc0688d-d611-4f20-85b3-9afcec9583ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.068116] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0675b5ec-0a19-441c-9950-78a2884a36e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.107668] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c37374b-67a2-4254-90f3-499b866c146d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.117620] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbec48c-dde1-4157-8c77-3db5df390a1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.121508] env[65121]: DEBUG nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 779.136171] env[65121]: DEBUG nova.compute.provider_tree [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.155085] env[65121]: DEBUG nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 779.271183] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.272273] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.464051] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106379, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.521708] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106382, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.644669] env[65121]: DEBUG nova.scheduler.client.report [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 779.650459] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.650816] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "e2d32a5e-c350-4b2b-9243-c3b412193a82" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.651021] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.651207] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "e2d32a5e-c350-4b2b-9243-c3b412193a82-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.651497] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.651596] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.653580] env[65121]: INFO nova.compute.manager [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Terminating instance [ 779.683473] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.952166] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "c1004635-b318-489c-9e16-6cb545279953" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.952536] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c1004635-b318-489c-9e16-6cb545279953" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.953247] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "c1004635-b318-489c-9e16-6cb545279953-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.953456] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c1004635-b318-489c-9e16-6cb545279953-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.953655] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c1004635-b318-489c-9e16-6cb545279953-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.957280] env[65121]: INFO nova.compute.manager [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Terminating instance [ 779.975030] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106379, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.027851] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106382, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.154271] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.157487] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.777s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.161021] env[65121]: INFO nova.compute.claims [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.163957] env[65121]: DEBUG nova.compute.manager [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 780.164234] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.165213] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866d3591-7cc2-4ff6-978c-3ebd15205e8b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.174366] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.174626] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39d8ee7e-0110-4532-9c02-94334b80a502 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.183054] env[65121]: DEBUG oslo_vmware.api [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 780.183054] env[65121]: value = "task-5106383" [ 780.183054] env[65121]: _type = "Task" [ 780.183054] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.196433] env[65121]: DEBUG oslo_vmware.api [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.202292] env[65121]: INFO nova.scheduler.client.report [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Deleted allocations for instance a43879c7-7378-4700-9581-8daabdafe5b2 [ 780.360036] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.360315] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.466356] env[65121]: DEBUG nova.compute.manager [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 780.467840] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.467840] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106379, 'name': CreateVM_Task, 'duration_secs': 1.576648} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.468651] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e37a956-8c27-44d9-9e42-a41733426ef2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.472035] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 780.475185] env[65121]: WARNING neutronclient.v2_0.client [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 780.475185] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.475185] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 780.475185] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 780.475185] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89b69e44-8c02-4834-a121-24407273c317 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.480828] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.481964] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6668ef58-44f5-45cd-bfec-491203f11e85 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.483882] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 780.483882] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c8d2d6-24ac-a14a-b3c9-4a761a5865ac" [ 780.483882] env[65121]: _type = "Task" [ 780.483882] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.491586] env[65121]: DEBUG oslo_vmware.api [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 780.491586] env[65121]: value = "task-5106384" [ 780.491586] env[65121]: _type = "Task" [ 780.491586] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.501115] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c8d2d6-24ac-a14a-b3c9-4a761a5865ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.507894] env[65121]: DEBUG oslo_vmware.api [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106384, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.522427] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106382, 'name': CreateVM_Task, 'duration_secs': 1.228428} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.522718] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 780.523544] env[65121]: WARNING neutronclient.v2_0.client [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 780.525109] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.626070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "75114f97-fe50-4624-9333-303e411529ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.627285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "75114f97-fe50-4624-9333-303e411529ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.627285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "75114f97-fe50-4624-9333-303e411529ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.627285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "75114f97-fe50-4624-9333-303e411529ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.627285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "75114f97-fe50-4624-9333-303e411529ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.629825] env[65121]: INFO nova.compute.manager [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Terminating instance [ 780.695607] env[65121]: DEBUG oslo_vmware.api [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106383, 'name': PowerOffVM_Task, 'duration_secs': 0.287662} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.696064] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.696271] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 780.696605] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b529c04-24b1-45d5-85ab-7673dde126b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.713435] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2e1aae7-3051-4879-adbc-f12f87516d0d tempest-ServersAdminNegativeTestJSON-1695887275 tempest-ServersAdminNegativeTestJSON-1695887275-project-member] Lock "a43879c7-7378-4700-9581-8daabdafe5b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.006s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.792871] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.793121] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.793308] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Deleting the datastore file [datastore2] e2d32a5e-c350-4b2b-9243-c3b412193a82 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.793584] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78841ca1-d984-458a-ac20-3da7d3ef7f5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.803928] env[65121]: DEBUG oslo_vmware.api [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 780.803928] env[65121]: value = "task-5106386" [ 780.803928] env[65121]: _type = "Task" [ 780.803928] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.816328] env[65121]: DEBUG oslo_vmware.api [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106386, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.000293] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c8d2d6-24ac-a14a-b3c9-4a761a5865ac, 'name': SearchDatastore_Task, 'duration_secs': 0.01467} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.003670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.003922] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.004454] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.004454] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.004454] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.005441] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.005745] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 781.005984] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6360b415-1643-466f-9cc6-52334126c07d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.013230] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea030598-788e-4d9c-878b-18ab44817300 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.021996] env[65121]: DEBUG oslo_vmware.api [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106384, 'name': PowerOffVM_Task, 'duration_secs': 0.281775} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.023240] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.023240] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.023240] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-669a92c6-37c0-4eb0-942c-703b432bb77a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.026660] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 781.026660] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5200fb5a-abe6-aa2d-966e-98b6ad4deda0" [ 781.026660] env[65121]: _type = "Task" [ 781.026660] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.039540] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5200fb5a-abe6-aa2d-966e-98b6ad4deda0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.041157] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.041397] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.042216] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22e4c223-662e-40bd-b640-7390502f29c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.050594] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 781.050594] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523bd816-ed69-95e9-1043-a7e0806572c9" [ 781.050594] env[65121]: _type = "Task" [ 781.050594] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.065913] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523bd816-ed69-95e9-1043-a7e0806572c9, 'name': SearchDatastore_Task, 'duration_secs': 0.012032} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.066842] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da0f7738-12b7-4e97-a492-a9e8edf79938 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.073823] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 781.073823] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52669c62-fc1f-b575-7bf9-e6344cdcf7d8" [ 781.073823] env[65121]: _type = "Task" [ 781.073823] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.084375] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52669c62-fc1f-b575-7bf9-e6344cdcf7d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.114265] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.114499] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.115369] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleting the datastore file [datastore2] c1004635-b318-489c-9e16-6cb545279953 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.115369] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86616184-ab40-4c4a-b3ed-425dca791836 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.124182] env[65121]: DEBUG oslo_vmware.api [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 781.124182] env[65121]: value = "task-5106388" [ 781.124182] env[65121]: _type = "Task" [ 781.124182] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.131896] env[65121]: DEBUG oslo_vmware.api [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.134834] env[65121]: DEBUG nova.compute.manager [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 781.135088] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.135906] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61e30c2-9a02-4622-ba6b-18a11a405214 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.145651] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 781.145651] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-659de42a-155a-4762-b376-b272eec7f564 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.152857] env[65121]: DEBUG oslo_vmware.api [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 781.152857] env[65121]: value = "task-5106389" [ 781.152857] env[65121]: _type = "Task" [ 781.152857] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.163542] env[65121]: DEBUG oslo_vmware.api [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106389, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.319185] env[65121]: DEBUG oslo_vmware.api [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106386, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27291} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.323133] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.323133] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.323133] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.323133] env[65121]: INFO nova.compute.manager [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Took 1.16 seconds to destroy the instance on the hypervisor. [ 781.323770] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 781.324420] env[65121]: DEBUG nova.compute.manager [-] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 781.324525] env[65121]: DEBUG nova.network.neutron [-] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 781.324792] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.325412] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.325702] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.489230] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.539469] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5200fb5a-abe6-aa2d-966e-98b6ad4deda0, 'name': SearchDatastore_Task, 'duration_secs': 0.026135} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.542855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.543200] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.543454] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.585770] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52669c62-fc1f-b575-7bf9-e6344cdcf7d8, 'name': SearchDatastore_Task, 'duration_secs': 0.01077} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.588912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.589202] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] b6f6871d-311c-4adb-824e-2907a12f4224/b6f6871d-311c-4adb-824e-2907a12f4224.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 781.589790] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.589979] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.590311] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0786d36-6ed1-4f15-8fa4-0e5968ee4c16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.592328] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-559d65e3-5ae3-428c-9de6-1e840be26596 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.605696] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 781.605696] env[65121]: value = "task-5106390" [ 781.605696] env[65121]: _type = "Task" [ 781.605696] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.605696] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.605696] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.606490] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fb5d369-acf3-4036-8a29-7d06ccedf903 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.622445] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 781.622445] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a151fe-3646-7c3c-1db2-65884a442095" [ 781.622445] env[65121]: _type = "Task" [ 781.622445] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.622951] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106390, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.637984] env[65121]: DEBUG oslo_vmware.api [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106388, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156232} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.638244] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a151fe-3646-7c3c-1db2-65884a442095, 'name': SearchDatastore_Task, 'duration_secs': 0.009984} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.641066] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.641261] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.641431] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.641625] env[65121]: INFO nova.compute.manager [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c1004635-b318-489c-9e16-6cb545279953] Took 1.17 seconds to destroy the instance on the hypervisor. [ 781.641829] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 781.643449] env[65121]: DEBUG nova.compute.manager [-] [instance: c1004635-b318-489c-9e16-6cb545279953] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 781.643449] env[65121]: DEBUG nova.network.neutron [-] [instance: c1004635-b318-489c-9e16-6cb545279953] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 781.643596] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.644031] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.644283] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.650991] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bd8a1a8-b5e4-4b2f-a200-3a90437cc4f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.662226] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 781.662226] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5209333f-3d65-49a9-a1e8-ef43e933b712" [ 781.662226] env[65121]: _type = "Task" [ 781.662226] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.668676] env[65121]: DEBUG oslo_vmware.api [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106389, 'name': PowerOffVM_Task, 'duration_secs': 0.230076} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.669416] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.669614] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.669927] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be6b7636-fed7-4de4-b2a8-e4fe5911306a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.677751] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5209333f-3d65-49a9-a1e8-ef43e933b712, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.752549] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.753069] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.753069] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleting the datastore file [datastore1] 75114f97-fe50-4624-9333-303e411529ea {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.753240] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2438ed5-be8a-465e-913f-048afb2c3ddc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.761218] env[65121]: DEBUG oslo_vmware.api [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 781.761218] env[65121]: value = "task-5106392" [ 781.761218] env[65121]: _type = "Task" [ 781.761218] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.771426] env[65121]: DEBUG oslo_vmware.api [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.780642] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3afc39-1025-4b8b-a204-05f716ff2db6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.789848] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c090e58a-e96e-45af-887c-c7837341b590 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.822751] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed2f86d-a5f4-4018-ae12-5844327cebf7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.833074] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b46f24a-acac-45d5-a397-640735ee1f01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.857833] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 782.119859] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106390, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.177514] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5209333f-3d65-49a9-a1e8-ef43e933b712, 'name': SearchDatastore_Task, 'duration_secs': 0.014248} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.177772] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.178051] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 00362477-c89e-4f60-98a4-d4928081d55e/00362477-c89e-4f60-98a4-d4928081d55e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 782.178323] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4600cf4d-10f6-4491-bce5-aa0e31f8b44c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.187012] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 782.187012] env[65121]: value = "task-5106393" [ 782.187012] env[65121]: _type = "Task" [ 782.187012] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.200600] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.274085] env[65121]: DEBUG oslo_vmware.api [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.44295} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.274369] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.274549] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.274880] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.275168] env[65121]: INFO nova.compute.manager [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 75114f97-fe50-4624-9333-303e411529ea] Took 1.14 seconds to destroy the instance on the hypervisor. [ 782.275458] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 782.275684] env[65121]: DEBUG nova.compute.manager [-] [instance: 75114f97-fe50-4624-9333-303e411529ea] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 782.275780] env[65121]: DEBUG nova.network.neutron [-] [instance: 75114f97-fe50-4624-9333-303e411529ea] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 782.276100] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.276667] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.276926] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.309927] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.384363] env[65121]: ERROR nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [req-838f1049-eade-4ce8-91e2-4022be9068c9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-838f1049-eade-4ce8-91e2-4022be9068c9"}]} [ 782.405793] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 782.423420] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 782.423863] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 782.441848] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 782.468787] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 782.629657] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106390, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544736} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.633679] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] b6f6871d-311c-4adb-824e-2907a12f4224/b6f6871d-311c-4adb-824e-2907a12f4224.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 782.633679] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.633679] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d80f7b2e-6ebc-4eac-8e0f-87b527e3e967 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.643874] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 782.643874] env[65121]: value = "task-5106394" [ 782.643874] env[65121]: _type = "Task" [ 782.643874] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.656941] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.703042] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106393, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.726636] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.143828] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0ef952-0b23-4f99-a7a6-b16880ed1fc6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.160108] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f21147-32cd-4cc9-b66e-dfa8b96a10a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.164088] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097312} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.164396] env[65121]: DEBUG nova.network.neutron [-] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.167327] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.167327] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0986ed8-07b8-45bc-a349-cbdfc4a053aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.206274] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af808fae-3948-4596-bdac-6a38b0e4f55a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.230153] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] b6f6871d-311c-4adb-824e-2907a12f4224/b6f6871d-311c-4adb-824e-2907a12f4224.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.232203] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecd93b53-af1c-4471-96b6-b2d7322fece8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.254600] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106393, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.785482} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.256286] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f442ca-38d9-4067-83e9-983e793d1982 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.261712] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 00362477-c89e-4f60-98a4-d4928081d55e/00362477-c89e-4f60-98a4-d4928081d55e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.261980] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.264108] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71b2e2b3-8876-4860-8da1-bf931c6aa60f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.267143] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 783.267143] env[65121]: value = "task-5106395" [ 783.267143] env[65121]: _type = "Task" [ 783.267143] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.280291] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.284706] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 783.284706] env[65121]: value = "task-5106396" [ 783.284706] env[65121]: _type = "Task" [ 783.284706] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.292426] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106395, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.298446] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106396, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.351143] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.352093] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.472348] env[65121]: DEBUG nova.network.neutron [-] [instance: c1004635-b318-489c-9e16-6cb545279953] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.673052] env[65121]: INFO nova.compute.manager [-] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Took 2.35 seconds to deallocate network for instance. [ 783.779898] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106395, 'name': ReconfigVM_Task, 'duration_secs': 0.361609} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.780315] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Reconfigured VM instance instance-0000002e to attach disk [datastore2] b6f6871d-311c-4adb-824e-2907a12f4224/b6f6871d-311c-4adb-824e-2907a12f4224.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.781489] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29253dd1-d38b-4d15-b3dd-ec166ca75be6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.796017] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 783.796017] env[65121]: value = "task-5106397" [ 783.796017] env[65121]: _type = "Task" [ 783.796017] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.804453] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079384} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.808372] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.809429] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106397, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.809515] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2e31ea-3344-406a-8869-f9f67c5a9c1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.814402] env[65121]: ERROR nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [req-be43aa44-fe38-4839-832f-70be4f9cda0b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-be43aa44-fe38-4839-832f-70be4f9cda0b"}]} [ 783.829141] env[65121]: DEBUG nova.network.neutron [-] [instance: 75114f97-fe50-4624-9333-303e411529ea] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.845326] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 00362477-c89e-4f60-98a4-d4928081d55e/00362477-c89e-4f60-98a4-d4928081d55e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.846969] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 783.852130] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a345348-7374-4113-8367-d16b4311f79e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.875183] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 783.875183] env[65121]: value = "task-5106398" [ 783.875183] env[65121]: _type = "Task" [ 783.875183] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.883153] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 783.883153] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.889011] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106398, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.898665] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 783.919862] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 783.975784] env[65121]: INFO nova.compute.manager [-] [instance: c1004635-b318-489c-9e16-6cb545279953] Took 2.33 seconds to deallocate network for instance. [ 784.151919] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Received event network-changed-00cda2d7-d8b7-4d99-9030-9f175852e4d9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 784.152453] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Refreshing instance network info cache due to event network-changed-00cda2d7-d8b7-4d99-9030-9f175852e4d9. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 784.152707] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Acquiring lock "refresh_cache-b6f6871d-311c-4adb-824e-2907a12f4224" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.152906] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Acquired lock "refresh_cache-b6f6871d-311c-4adb-824e-2907a12f4224" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 784.153130] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Refreshing network info cache for port 00cda2d7-d8b7-4d99-9030-9f175852e4d9 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 784.181904] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.310191] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106397, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.348755] env[65121]: INFO nova.compute.manager [-] [instance: 75114f97-fe50-4624-9333-303e411529ea] Took 2.07 seconds to deallocate network for instance. [ 784.392014] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.483758] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.553641] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e15906-e039-4156-bc52-bd3ab586dd4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.564332] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a317d0-8639-44da-9347-24aed927fd33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.599328] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a161583-0aee-4079-8ddd-7dbc626afd1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.608038] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5802e7c7-c1f9-4c8e-bf8e-6e6efa1db14b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.625187] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.661030] env[65121]: WARNING neutronclient.v2_0.client [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 784.661304] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 784.661685] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 784.812196] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106397, 'name': Rename_Task, 'duration_secs': 0.909883} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.812374] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 784.812647] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dceec7ab-85e9-445a-b530-82fc521b47c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.821495] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 784.821495] env[65121]: value = "task-5106399" [ 784.821495] env[65121]: _type = "Task" [ 784.821495] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.828402] env[65121]: DEBUG nova.compute.manager [req-7f756296-9c4e-4d46-8aee-5715cfa9ece2 req-b22e8717-6d83-46bc-ac65-1e9834b20261 service nova] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Received event network-vif-deleted-b25aa25b-2d1b-414e-883a-324894d4b483 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 784.832111] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106399, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.857586] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.891438] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106398, 'name': ReconfigVM_Task, 'duration_secs': 0.978895} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.891711] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 00362477-c89e-4f60-98a4-d4928081d55e/00362477-c89e-4f60-98a4-d4928081d55e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.892473] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c70c5218-e503-43f4-8dfe-0322e01804d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.900640] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 784.900640] env[65121]: value = "task-5106400" [ 784.900640] env[65121]: _type = "Task" [ 784.900640] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.910905] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106400, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.162449] env[65121]: ERROR nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [req-4ac8423e-9c3a-45d6-9db6-bad004e5ce0c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4ac8423e-9c3a-45d6-9db6-bad004e5ce0c"}]} [ 785.178517] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "aab9317b-4ee6-48b3-905b-859a5996f33d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.178975] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.178975] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "aab9317b-4ee6-48b3-905b-859a5996f33d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.179796] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.179796] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.182223] env[65121]: INFO nova.compute.manager [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Terminating instance [ 785.184684] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 785.205954] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 785.205954] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 785.222256] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 785.251375] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 785.335419] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106399, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.411753] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106400, 'name': Rename_Task, 'duration_secs': 0.197539} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.412748] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 785.413046] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90c4a862-add6-4ffe-9e79-88f83ad88fd4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.423991] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 785.423991] env[65121]: value = "task-5106401" [ 785.423991] env[65121]: _type = "Task" [ 785.423991] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.434559] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106401, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.490708] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.491290] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.690166] env[65121]: DEBUG nova.compute.manager [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 785.690166] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.690595] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ed0db5-1d4c-4404-82fd-1ffbbb11cc8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.703565] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 785.703862] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be4db056-dfa2-49d3-8f8f-1a6b2041fac1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.719941] env[65121]: DEBUG oslo_vmware.api [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 785.719941] env[65121]: value = "task-5106402" [ 785.719941] env[65121]: _type = "Task" [ 785.719941] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.723934] env[65121]: DEBUG nova.compute.manager [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 785.740985] env[65121]: DEBUG oslo_vmware.api [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106402, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.834235] env[65121]: DEBUG oslo_vmware.api [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106399, 'name': PowerOnVM_Task, 'duration_secs': 0.538801} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.838033] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 785.838033] env[65121]: INFO nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Took 13.15 seconds to spawn the instance on the hypervisor. [ 785.838033] env[65121]: DEBUG nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 785.838033] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fd496f-5103-4b11-9fc4-d0664e13a501 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.890615] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "19b1f3f9-842e-4150-8890-b0b22393c3af" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.890615] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.890615] env[65121]: INFO nova.compute.manager [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Rebooting instance [ 785.940529] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106401, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.941772] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a11ce3d-7510-450f-a0e7-7b9a43f47d4a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.950167] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87eb6311-6597-47b5-b46a-67a831ac8898 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.989693] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3383ec97-a1d9-4c3d-b17c-2d4a0603f953 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.999670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd655b4-0115-4d77-a949-aab196acf8e9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.019723] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.124936] env[65121]: WARNING neutronclient.v2_0.client [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.125700] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.126212] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.235795] env[65121]: DEBUG oslo_vmware.api [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106402, 'name': PowerOffVM_Task, 'duration_secs': 0.233239} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.236805] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.236805] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 786.237846] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40a9c4a3-72ec-49bc-8cc0-b8b55b1fecbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.252669] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.324356] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 786.324356] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 786.324356] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Deleting the datastore file [datastore1] aab9317b-4ee6-48b3-905b-859a5996f33d {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.324356] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63384164-64a2-4bb2-9a54-26efc32a3ded {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.334517] env[65121]: DEBUG oslo_vmware.api [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for the task: (returnval){ [ 786.334517] env[65121]: value = "task-5106404" [ 786.334517] env[65121]: _type = "Task" [ 786.334517] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.341506] env[65121]: DEBUG oslo_vmware.api [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.360926] env[65121]: INFO nova.compute.manager [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Took 49.40 seconds to build instance. [ 786.413851] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.414049] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquired lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.415185] env[65121]: DEBUG nova.network.neutron [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 786.434634] env[65121]: DEBUG oslo_vmware.api [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106401, 'name': PowerOnVM_Task, 'duration_secs': 0.86516} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.435152] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 786.435240] env[65121]: INFO nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Took 10.46 seconds to spawn the instance on the hypervisor. [ 786.435360] env[65121]: DEBUG nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 786.436186] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d13174-b19e-4fab-9513-f122b642724f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.569019] env[65121]: DEBUG nova.scheduler.client.report [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 74 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 786.569019] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 74 to 75 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 786.569019] env[65121]: DEBUG nova.compute.provider_tree [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 786.851058] env[65121]: DEBUG oslo_vmware.api [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Task: {'id': task-5106404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159617} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.851313] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.851501] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 786.851695] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.851967] env[65121]: INFO nova.compute.manager [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 786.852305] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 786.853089] env[65121]: DEBUG nova.compute.manager [-] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 786.853089] env[65121]: DEBUG nova.network.neutron [-] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 786.853089] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.854081] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.854081] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.863891] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c55c9b5e-dced-4027-b1a5-b81fb4a81097 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "b6f6871d-311c-4adb-824e-2907a12f4224" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.906s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.916697] env[65121]: WARNING neutronclient.v2_0.client [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.917441] env[65121]: WARNING openstack [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.917781] env[65121]: WARNING openstack [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.958917] env[65121]: INFO nova.compute.manager [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Took 42.54 seconds to build instance. [ 787.069643] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Updated VIF entry in instance network info cache for port 00cda2d7-d8b7-4d99-9030-9f175852e4d9. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 787.070231] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Updating instance_info_cache with network_info: [{"id": "00cda2d7-d8b7-4d99-9030-9f175852e4d9", "address": "fa:16:3e:55:35:54", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00cda2d7-d8", "ovs_interfaceid": "00cda2d7-d8b7-4d99-9030-9f175852e4d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.074783] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 6.918s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.075393] env[65121]: DEBUG nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 787.077958] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.591s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.079444] env[65121]: INFO nova.compute.claims [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.104507] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.367224] env[65121]: DEBUG nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 787.460893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d12dc8a-1d55-480a-a036-2c5cbe8da8dd tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "00362477-c89e-4f60-98a4-d4928081d55e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.528s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.573814] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Releasing lock "refresh_cache-b6f6871d-311c-4adb-824e-2907a12f4224" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.574235] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Received event network-vif-plugged-7381b4f6-e691-437c-9b26-59e3192fe6ed {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 787.574662] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Acquiring lock "00362477-c89e-4f60-98a4-d4928081d55e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.574731] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Lock "00362477-c89e-4f60-98a4-d4928081d55e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.575594] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Lock "00362477-c89e-4f60-98a4-d4928081d55e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.575594] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] No waiting events found dispatching network-vif-plugged-7381b4f6-e691-437c-9b26-59e3192fe6ed {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 787.575594] env[65121]: WARNING nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Received unexpected event network-vif-plugged-7381b4f6-e691-437c-9b26-59e3192fe6ed for instance with vm_state building and task_state spawning. [ 787.575879] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Received event network-changed-7381b4f6-e691-437c-9b26-59e3192fe6ed {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 787.576320] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Refreshing instance network info cache due to event network-changed-7381b4f6-e691-437c-9b26-59e3192fe6ed. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 787.576320] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Acquiring lock "refresh_cache-00362477-c89e-4f60-98a4-d4928081d55e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.576576] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Acquired lock "refresh_cache-00362477-c89e-4f60-98a4-d4928081d55e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 787.577690] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Refreshing network info cache for port 7381b4f6-e691-437c-9b26-59e3192fe6ed {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 787.586446] env[65121]: DEBUG nova.compute.utils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 787.590662] env[65121]: DEBUG nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 787.590662] env[65121]: DEBUG nova.network.neutron [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 787.590797] env[65121]: WARNING neutronclient.v2_0.client [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.591098] env[65121]: WARNING neutronclient.v2_0.client [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.595109] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.595109] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.796725] env[65121]: DEBUG nova.policy [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a602d2a28164f069851eb46c57f47ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '693c1c83a1d5453bbedaf8ce2bf7c8d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 787.899892] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.956490] env[65121]: DEBUG nova.network.neutron [-] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.964269] env[65121]: DEBUG nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 788.083344] env[65121]: WARNING neutronclient.v2_0.client [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.084462] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.084945] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.105900] env[65121]: DEBUG nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 788.438731] env[65121]: WARNING openstack [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.440320] env[65121]: WARNING openstack [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.462130] env[65121]: INFO nova.compute.manager [-] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Took 1.61 seconds to deallocate network for instance. [ 788.492845] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.616568] env[65121]: DEBUG nova.network.neutron [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Successfully created port: 1740aa3b-2428-4bc3-8888-c4edc0734c60 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 788.781910] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189053dd-29c0-44dd-b468-d41cdaace450 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.791017] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305c00a9-2e03-4ca0-8f9b-1bade260ef1c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.821801] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a19df1-fca2-4a62-b029-0056fb903286 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.830073] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa8b0e2-f6b0-4a70-b0de-ceead1887466 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.844307] env[65121]: DEBUG nova.compute.provider_tree [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.979374] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 789.122033] env[65121]: DEBUG nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 789.163863] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 789.164242] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 789.164454] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 789.164590] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 789.164728] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 789.164903] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 789.165223] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.165391] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 789.165578] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 789.165738] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 789.165971] env[65121]: DEBUG nova.virt.hardware [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 789.166865] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5889b70f-8579-451a-aad8-649c70eceed8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.176013] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1faa8e30-7593-4704-813c-6f70ae976d98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.347837] env[65121]: DEBUG nova.scheduler.client.report [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 789.436503] env[65121]: WARNING neutronclient.v2_0.client [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.437264] env[65121]: WARNING openstack [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.438023] env[65121]: WARNING openstack [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.464478] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.464478] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.853929] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.776s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.854464] env[65121]: DEBUG nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 789.857369] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.244s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.858786] env[65121]: INFO nova.compute.claims [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.245186] env[65121]: DEBUG nova.network.neutron [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Successfully updated port: 1740aa3b-2428-4bc3-8888-c4edc0734c60 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 790.365711] env[65121]: DEBUG nova.compute.utils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 790.369100] env[65121]: DEBUG nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 790.369100] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 790.369100] env[65121]: WARNING neutronclient.v2_0.client [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.369100] env[65121]: WARNING neutronclient.v2_0.client [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.369331] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.369778] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.464640] env[65121]: WARNING neutronclient.v2_0.client [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.468512] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.468512] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.483175] env[65121]: DEBUG nova.network.neutron [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updating instance_info_cache with network_info: [{"id": "cd096b57-4e55-46f3-b323-cc951f405db2", "address": "fa:16:3e:40:1f:eb", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd096b57-4e", "ovs_interfaceid": "cd096b57-4e55-46f3-b323-cc951f405db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.509129] env[65121]: DEBUG nova.policy [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e8520cf91b64499b55cebfca04cec5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc44d2096bff489193193973fe3a3550', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 790.563551] env[65121]: DEBUG nova.compute.manager [req-da0777d5-db7a-4277-90a7-3a9debf101b6 req-7ff24a04-ae4a-414b-a28e-1a789614b1df service nova] [instance: 75114f97-fe50-4624-9333-303e411529ea] Received event network-vif-deleted-d00d6f6e-2d62-4ae9-8df3-0a5f04428080 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 790.612735] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Updated VIF entry in instance network info cache for port 7381b4f6-e691-437c-9b26-59e3192fe6ed. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 790.613183] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Updating instance_info_cache with network_info: [{"id": "7381b4f6-e691-437c-9b26-59e3192fe6ed", "address": "fa:16:3e:3d:a9:02", "network": {"id": "e9a0f265-bbfd-40d4-a9b1-63fbae017794", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-58006937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f6094b0e36ba4861949628aae919f102", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7836a5b-a91e-4d3f-8e96-afe024f62bb5", "external-id": "nsx-vlan-transportzone-419", "segmentation_id": 419, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7381b4f6-e6", "ovs_interfaceid": "7381b4f6-e691-437c-9b26-59e3192fe6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.749584] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.749806] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.750231] env[65121]: DEBUG nova.network.neutron [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 790.797902] env[65121]: DEBUG nova.compute.manager [req-19ba6526-c799-4a3c-bc04-fdab92293c81 req-af9baa9c-5021-4248-aba5-b34712bd8d06 service nova] [instance: c1004635-b318-489c-9e16-6cb545279953] Received event network-vif-deleted-2cd8c718-cf4c-4d35-92ae-d95f45242770 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 790.839753] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Successfully created port: ac5e737f-9b34-4d3a-896b-163829980cce {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 790.879058] env[65121]: DEBUG nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 790.987649] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Releasing lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.121536] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Releasing lock "refresh_cache-00362477-c89e-4f60-98a4-d4928081d55e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.121805] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Received event network-changed-cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 791.121939] env[65121]: DEBUG nova.compute.manager [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Refreshing instance network info cache due to event network-changed-cd096b57-4e55-46f3-b323-cc951f405db2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 791.122190] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Acquiring lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.122289] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Acquired lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 791.122438] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Refreshing network info cache for port cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 791.252466] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Successfully created port: 183f83c8-107c-47b0-9081-d9a9173bc9b3 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 791.255845] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.256440] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.326411] env[65121]: DEBUG nova.network.neutron [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 791.399131] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.399504] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.466359] env[65121]: WARNING neutronclient.v2_0.client [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 791.467080] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.467432] env[65121]: WARNING openstack [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.496937] env[65121]: DEBUG nova.compute.manager [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 791.497223] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e529090-bca4-4ae5-8e5e-cc9b2e7f9683 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.594998] env[65121]: DEBUG nova.network.neutron [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Updating instance_info_cache with network_info: [{"id": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "address": "fa:16:3e:81:57:23", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1740aa3b-24", "ovs_interfaceid": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 791.630009] env[65121]: WARNING neutronclient.v2_0.client [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 791.630709] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.631058] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.641872] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1384e63-ccaa-4bed-b123-d0d02b276b44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.652985] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab49b97e-dd32-4269-8b34-be4a08a52f16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.689461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7757c5f-00be-4ae5-9f46-2e4d8d6f6ee4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.698222] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2bc5702-d037-442b-8511-a0b2e34a3cf6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.711894] env[65121]: DEBUG nova.compute.provider_tree [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.878484] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.878931] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.893138] env[65121]: DEBUG nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 791.929271] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 791.929516] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 791.929664] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 791.929841] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 791.930055] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 791.930508] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 791.930652] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.930903] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 791.931151] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 791.931375] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 791.931684] env[65121]: DEBUG nova.virt.hardware [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 791.932741] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15759942-58e6-4543-8d80-89d700983295 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.942822] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3ba258-5a27-4aad-a393-f7161e82d011 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.980581] env[65121]: WARNING neutronclient.v2_0.client [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 791.980979] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.981247] env[65121]: WARNING openstack [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.097590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 792.098033] env[65121]: DEBUG nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Instance network_info: |[{"id": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "address": "fa:16:3e:81:57:23", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1740aa3b-24", "ovs_interfaceid": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 792.098530] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:57:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1740aa3b-2428-4bc3-8888-c4edc0734c60', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.106762] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Creating folder: Project (693c1c83a1d5453bbedaf8ce2bf7c8d1). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.107514] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-533cd8fa-89cc-40a6-a0c6-fab9f7392677 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.124224] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Created folder: Project (693c1c83a1d5453bbedaf8ce2bf7c8d1) in parent group-v993268. [ 792.124434] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Creating folder: Instances. Parent ref: group-v993404. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.124709] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fde19b50-3b0b-4367-bef2-11a3a3a5b0a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.136567] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Created folder: Instances in parent group-v993404. [ 792.138065] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 792.138065] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.138065] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7557eec6-8454-462f-b172-b891d7c26c08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.159888] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.159888] env[65121]: value = "task-5106407" [ 792.159888] env[65121]: _type = "Task" [ 792.159888] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.169371] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106407, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.207141] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "5d352184-7074-4fab-b4ac-cc11a6b936b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.207376] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.211345] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updated VIF entry in instance network info cache for port cd096b57-4e55-46f3-b323-cc951f405db2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 792.211416] env[65121]: DEBUG nova.network.neutron [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updating instance_info_cache with network_info: [{"id": "cd096b57-4e55-46f3-b323-cc951f405db2", "address": "fa:16:3e:40:1f:eb", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd096b57-4e", "ovs_interfaceid": "cd096b57-4e55-46f3-b323-cc951f405db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 792.217937] env[65121]: DEBUG nova.scheduler.client.report [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 792.525738] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361f7ba8-51a2-4009-adfa-87b91916a99c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.535731] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Doing hard reboot of VM {{(pid=65121) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 792.536083] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-7fa1226d-9fb2-4094-a0dc-33aeebc5e0f1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.543998] env[65121]: DEBUG oslo_vmware.api [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 792.543998] env[65121]: value = "task-5106408" [ 792.543998] env[65121]: _type = "Task" [ 792.543998] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.556300] env[65121]: DEBUG oslo_vmware.api [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106408, 'name': ResetVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.669673] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106407, 'name': CreateVM_Task, 'duration_secs': 0.398307} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.669850] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 792.670422] env[65121]: WARNING neutronclient.v2_0.client [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.670788] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.670937] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.671296] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 792.671590] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbc43d7b-0c01-48e8-874e-defb73cb1630 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.677224] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 792.677224] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b03b0c-0120-985a-b362-80c5d34d98a4" [ 792.677224] env[65121]: _type = "Task" [ 792.677224] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.690786] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b03b0c-0120-985a-b362-80c5d34d98a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.715680] env[65121]: DEBUG oslo_concurrency.lockutils [req-ae31c802-530b-4476-9324-649b44430239 req-b1a116b4-aeb9-4f96-9f9b-be0d7e7e7068 service nova] Releasing lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 792.724930] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.868s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.725444] env[65121]: DEBUG nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 792.728828] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.698s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.729108] env[65121]: DEBUG nova.objects.instance [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lazy-loading 'resources' on Instance uuid 4b6fe349-335e-4202-ab76-04f6ccc036ee {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.897197] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Successfully updated port: ac5e737f-9b34-4d3a-896b-163829980cce {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 793.048232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.048232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.060181] env[65121]: DEBUG oslo_vmware.api [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106408, 'name': ResetVM_Task, 'duration_secs': 0.10871} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.062762] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Did hard reboot of VM {{(pid=65121) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 793.062762] env[65121]: DEBUG nova.compute.manager [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 793.062762] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845e8fad-4490-4158-a5bf-79aeaa0f277d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.189983] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b03b0c-0120-985a-b362-80c5d34d98a4, 'name': SearchDatastore_Task, 'duration_secs': 0.028358} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.190166] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.190438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.190614] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.190754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.191390] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.191390] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7de3ffe3-9c58-465b-b025-4909fc546670 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.203522] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.203725] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.204721] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d67235f-dc99-410d-835a-1f1310a9d64b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.214734] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 793.214734] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528fa37b-1f2b-7307-b841-054a20cab6cd" [ 793.214734] env[65121]: _type = "Task" [ 793.214734] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.223025] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528fa37b-1f2b-7307-b841-054a20cab6cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.240715] env[65121]: DEBUG nova.compute.utils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 793.241943] env[65121]: DEBUG nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 793.242526] env[65121]: DEBUG nova.network.neutron [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 793.243153] env[65121]: WARNING neutronclient.v2_0.client [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 793.243877] env[65121]: WARNING neutronclient.v2_0.client [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 793.245059] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.246393] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.348552] env[65121]: DEBUG nova.policy [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8931a266fddc4346b0c5243608a39bff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00383aa3355e438cb703c2b86c7917f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 793.554462] env[65121]: DEBUG nova.compute.manager [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Received event network-vif-plugged-1740aa3b-2428-4bc3-8888-c4edc0734c60 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 793.554652] env[65121]: DEBUG oslo_concurrency.lockutils [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Acquiring lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.554746] env[65121]: DEBUG oslo_concurrency.lockutils [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.554960] env[65121]: DEBUG oslo_concurrency.lockutils [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.555143] env[65121]: DEBUG nova.compute.manager [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] No waiting events found dispatching network-vif-plugged-1740aa3b-2428-4bc3-8888-c4edc0734c60 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 793.555300] env[65121]: WARNING nova.compute.manager [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Received unexpected event network-vif-plugged-1740aa3b-2428-4bc3-8888-c4edc0734c60 for instance with vm_state building and task_state spawning. [ 793.555466] env[65121]: DEBUG nova.compute.manager [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Received event network-changed-1740aa3b-2428-4bc3-8888-c4edc0734c60 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 793.555621] env[65121]: DEBUG nova.compute.manager [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Refreshing instance network info cache due to event network-changed-1740aa3b-2428-4bc3-8888-c4edc0734c60. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 793.556192] env[65121]: DEBUG oslo_concurrency.lockutils [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Acquiring lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.556192] env[65121]: DEBUG oslo_concurrency.lockutils [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Acquired lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.556192] env[65121]: DEBUG nova.network.neutron [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Refreshing network info cache for port 1740aa3b-2428-4bc3-8888-c4edc0734c60 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 793.577103] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee09d664-7da6-4d90-8293-7d8743e80810 tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 7.687s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.727020] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528fa37b-1f2b-7307-b841-054a20cab6cd, 'name': SearchDatastore_Task, 'duration_secs': 0.010472} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.727843] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21e0f49a-9495-4f7c-96a3-25addfd1fc60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.734866] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 793.734866] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ca6000-8293-3bd3-5dbd-5c6407e37341" [ 793.734866] env[65121]: _type = "Task" [ 793.734866] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.743974] env[65121]: DEBUG nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 793.755056] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ca6000-8293-3bd3-5dbd-5c6407e37341, 'name': SearchDatastore_Task, 'duration_secs': 0.011034} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.758202] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.759117] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/ab41ae41-e69d-47fb-a31a-16748fd3b0f2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 793.759712] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-459a1ae3-dc97-4289-afaf-bbd4e88f3f6d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.768171] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 793.768171] env[65121]: value = "task-5106409" [ 793.768171] env[65121]: _type = "Task" [ 793.768171] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.782105] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106409, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.865296] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ad5af5-7955-46d6-aabc-82afb9c46307 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.874206] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff88d45-0bbf-4bd7-8903-d99489351afc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.879180] env[65121]: DEBUG nova.network.neutron [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Successfully created port: 4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 793.908594] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad133d55-5231-43f4-9920-81fb2956d33d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.919554] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64f4874-1ea1-4497-a6a8-2226c393b87f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.927221] env[65121]: DEBUG nova.compute.manager [req-70e1e066-b1dd-4505-8389-1d8c2c6ac1c8 req-e68f091b-4f91-41cb-97ed-4a04746c406c service nova] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Received event network-vif-deleted-4dd0a4a1-30fb-4c71-b7d8-e51204984bd4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 793.940324] env[65121]: DEBUG nova.compute.provider_tree [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.059609] env[65121]: WARNING neutronclient.v2_0.client [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.060588] env[65121]: WARNING openstack [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.061062] env[65121]: WARNING openstack [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.283551] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106409, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499231} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.283551] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/ab41ae41-e69d-47fb-a31a-16748fd3b0f2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 794.283551] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.283551] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2ec1ce1-b5a2-48d7-9bfd-82bfd047b79d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.294258] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 794.294258] env[65121]: value = "task-5106410" [ 794.294258] env[65121]: _type = "Task" [ 794.294258] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.308570] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106410, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.443413] env[65121]: DEBUG nova.scheduler.client.report [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 794.759367] env[65121]: DEBUG nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 794.783665] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 794.783938] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 794.784058] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 794.784218] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 794.784370] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 794.784486] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 794.784686] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.784966] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 794.785167] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 794.785329] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 794.785506] env[65121]: DEBUG nova.virt.hardware [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 794.786473] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b7d8ea-6547-4372-a209-95f128678b47 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.795504] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff2e808-1231-4e71-a7a9-1237d8741a96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.816980] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106410, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083347} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.817087] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.817875] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd7638b-5e8d-4726-802b-c60c62d94b9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.842741] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/ab41ae41-e69d-47fb-a31a-16748fd3b0f2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.842913] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06e49692-3b35-4b43-983d-ba8ffb14bc06 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.864462] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 794.864462] env[65121]: value = "task-5106414" [ 794.864462] env[65121]: _type = "Task" [ 794.864462] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.873673] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106414, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.949043] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.220s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.951726] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.086s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 794.953347] env[65121]: INFO nova.compute.claims [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.973188] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Successfully updated port: 183f83c8-107c-47b0-9081-d9a9173bc9b3 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 794.982297] env[65121]: INFO nova.scheduler.client.report [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Deleted allocations for instance 4b6fe349-335e-4202-ab76-04f6ccc036ee [ 795.202995] env[65121]: WARNING openstack [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.203405] env[65121]: WARNING openstack [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.261364] env[65121]: WARNING neutronclient.v2_0.client [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.262088] env[65121]: WARNING openstack [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.262421] env[65121]: WARNING openstack [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.374908] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106414, 'name': ReconfigVM_Task, 'duration_secs': 0.340328} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.375240] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Reconfigured VM instance instance-00000030 to attach disk [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/ab41ae41-e69d-47fb-a31a-16748fd3b0f2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.375901] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d35b924a-dbd9-472e-aa92-e79f1ea2d2dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.384172] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 795.384172] env[65121]: value = "task-5106415" [ 795.384172] env[65121]: _type = "Task" [ 795.384172] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.393844] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106415, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.408413] env[65121]: DEBUG nova.network.neutron [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Updated VIF entry in instance network info cache for port 1740aa3b-2428-4bc3-8888-c4edc0734c60. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 795.408769] env[65121]: DEBUG nova.network.neutron [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Updating instance_info_cache with network_info: [{"id": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "address": "fa:16:3e:81:57:23", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1740aa3b-24", "ovs_interfaceid": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.476275] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.476499] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 795.477438] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 795.490588] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ace59ba-b458-43fd-82ce-31ac9c546103 tempest-ServersTestFqdnHostnames-521639732 tempest-ServersTestFqdnHostnames-521639732-project-member] Lock "4b6fe349-335e-4202-ab76-04f6ccc036ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.467s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.528468] env[65121]: DEBUG nova.network.neutron [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Successfully updated port: 4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 795.895488] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106415, 'name': Rename_Task, 'duration_secs': 0.37752} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.896391] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.896391] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ce683de-c492-41c3-8f79-1ff6215f7575 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.904537] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 795.904537] env[65121]: value = "task-5106416" [ 795.904537] env[65121]: _type = "Task" [ 795.904537] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.911950] env[65121]: DEBUG oslo_concurrency.lockutils [req-83322bcd-a302-4fb7-b38e-e43ff036a1df req-9475d30c-83f8-48bd-8aa2-a3ff8a829653 service nova] Releasing lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.912441] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.981277] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.981676] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.028082] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 796.030766] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.030963] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 796.031656] env[65121]: DEBUG nova.network.neutron [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 796.060709] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.061197] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.105468] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received event network-vif-plugged-ac5e737f-9b34-4d3a-896b-163829980cce {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 796.105721] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Acquiring lock "c1632a96-f334-4978-9f31-97f9329577e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.106049] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Lock "c1632a96-f334-4978-9f31-97f9329577e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.106265] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Lock "c1632a96-f334-4978-9f31-97f9329577e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.106418] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] No waiting events found dispatching network-vif-plugged-ac5e737f-9b34-4d3a-896b-163829980cce {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 796.106616] env[65121]: WARNING nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received unexpected event network-vif-plugged-ac5e737f-9b34-4d3a-896b-163829980cce for instance with vm_state building and task_state spawning. [ 796.106751] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received event network-changed-ac5e737f-9b34-4d3a-896b-163829980cce {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 796.106871] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Refreshing instance network info cache due to event network-changed-ac5e737f-9b34-4d3a-896b-163829980cce. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 796.108129] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Acquiring lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.183256] env[65121]: WARNING neutronclient.v2_0.client [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 796.184017] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.185292] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.199883] env[65121]: DEBUG nova.compute.manager [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received event network-vif-plugged-183f83c8-107c-47b0-9081-d9a9173bc9b3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 796.200159] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Acquiring lock "c1632a96-f334-4978-9f31-97f9329577e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.200415] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Lock "c1632a96-f334-4978-9f31-97f9329577e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.200613] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Lock "c1632a96-f334-4978-9f31-97f9329577e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.200815] env[65121]: DEBUG nova.compute.manager [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] No waiting events found dispatching network-vif-plugged-183f83c8-107c-47b0-9081-d9a9173bc9b3 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 796.201029] env[65121]: WARNING nova.compute.manager [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received unexpected event network-vif-plugged-183f83c8-107c-47b0-9081-d9a9173bc9b3 for instance with vm_state building and task_state spawning. [ 796.201215] env[65121]: DEBUG nova.compute.manager [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received event network-changed-183f83c8-107c-47b0-9081-d9a9173bc9b3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 796.201398] env[65121]: DEBUG nova.compute.manager [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Refreshing instance network info cache due to event network-changed-183f83c8-107c-47b0-9081-d9a9173bc9b3. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 796.201586] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Acquiring lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.416436] env[65121]: DEBUG oslo_vmware.api [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106416, 'name': PowerOnVM_Task, 'duration_secs': 0.49949} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.420716] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 796.420926] env[65121]: INFO nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Took 7.30 seconds to spawn the instance on the hypervisor. [ 796.421130] env[65121]: DEBUG nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 796.424255] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47c7eca-dce1-4cec-b661-cf01c7754ad9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.494233] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.494233] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.504685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "00362477-c89e-4f60-98a4-d4928081d55e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.504986] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "00362477-c89e-4f60-98a4-d4928081d55e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.505217] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "00362477-c89e-4f60-98a4-d4928081d55e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.505394] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "00362477-c89e-4f60-98a4-d4928081d55e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.505554] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "00362477-c89e-4f60-98a4-d4928081d55e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.507970] env[65121]: INFO nova.compute.manager [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Terminating instance [ 796.534221] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.534668] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.639600] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cd8084-3bb0-45b2-be6f-a3d22f9f1d95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.648990] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ed697e-b766-400f-b773-70a0be17a4ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.684180] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b607bc-d529-4a1c-b08a-69477445d499 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.693252] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d7e606-1605-4f1e-9f0a-f661fac9fe66 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.711519] env[65121]: DEBUG nova.compute.provider_tree [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.781878] env[65121]: DEBUG nova.network.neutron [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 796.839276] env[65121]: WARNING neutronclient.v2_0.client [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 796.840140] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.840551] env[65121]: WARNING openstack [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.860841] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.860899] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.943810] env[65121]: INFO nova.compute.manager [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Took 40.58 seconds to build instance. [ 797.012876] env[65121]: DEBUG nova.compute.manager [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 797.013105] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.014141] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ef7c0d-2091-45ed-b20c-6e4eec96f4ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.023335] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.023587] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-279e37fc-cdd2-422e-8c40-b6c97a8051dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.031048] env[65121]: DEBUG oslo_vmware.api [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 797.031048] env[65121]: value = "task-5106417" [ 797.031048] env[65121]: _type = "Task" [ 797.031048] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.035361] env[65121]: DEBUG nova.network.neutron [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Updating instance_info_cache with network_info: [{"id": "ac5e737f-9b34-4d3a-896b-163829980cce", "address": "fa:16:3e:83:cc:41", "network": {"id": "b62e3e62-2227-4ef5-b7db-529e77bc8e38", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-28468943", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.222", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15538852-1a3f-4f71-b4a9-4923c5837c4f", "external-id": "nsx-vlan-transportzone-165", "segmentation_id": 165, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac5e737f-9b", "ovs_interfaceid": "ac5e737f-9b34-4d3a-896b-163829980cce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "address": "fa:16:3e:2a:04:d8", "network": {"id": "92b9c32c-2ead-47d9-9c56-0065346f5d96", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-698410639", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41278529-8bd2-44a1-97c8-03967faa3ff7", "external-id": "nsx-vlan-transportzone-749", "segmentation_id": 749, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap183f83c8-10", "ovs_interfaceid": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 797.047356] env[65121]: DEBUG oslo_vmware.api [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.110844] env[65121]: WARNING neutronclient.v2_0.client [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 797.112284] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.112696] env[65121]: WARNING openstack [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.244803] env[65121]: ERROR nova.scheduler.client.report [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [req-8fe90d1a-1ebc-42a4-b3e1-bf68c39fff47] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8fe90d1a-1ebc-42a4-b3e1-bf68c39fff47"}]} [ 797.270504] env[65121]: DEBUG nova.scheduler.client.report [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 797.288463] env[65121]: DEBUG nova.scheduler.client.report [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 797.288940] env[65121]: DEBUG nova.compute.provider_tree [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.307373] env[65121]: DEBUG nova.scheduler.client.report [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 797.314092] env[65121]: DEBUG nova.network.neutron [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updating instance_info_cache with network_info: [{"id": "4304bec2-f452-4be6-aeda-3ad8354a4547", "address": "fa:16:3e:71:b3:4e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4304bec2-f4", "ovs_interfaceid": "4304bec2-f452-4be6-aeda-3ad8354a4547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 797.342091] env[65121]: DEBUG nova.scheduler.client.report [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 797.449215] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e5c64983-b451-4ef8-9854-d0757e0478cc tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.422s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.545502] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Releasing lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 797.545839] env[65121]: DEBUG nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Instance network_info: |[{"id": "ac5e737f-9b34-4d3a-896b-163829980cce", "address": "fa:16:3e:83:cc:41", "network": {"id": "b62e3e62-2227-4ef5-b7db-529e77bc8e38", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-28468943", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.222", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15538852-1a3f-4f71-b4a9-4923c5837c4f", "external-id": "nsx-vlan-transportzone-165", "segmentation_id": 165, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac5e737f-9b", "ovs_interfaceid": "ac5e737f-9b34-4d3a-896b-163829980cce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "address": "fa:16:3e:2a:04:d8", "network": {"id": "92b9c32c-2ead-47d9-9c56-0065346f5d96", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-698410639", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41278529-8bd2-44a1-97c8-03967faa3ff7", "external-id": "nsx-vlan-transportzone-749", "segmentation_id": 749, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap183f83c8-10", "ovs_interfaceid": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 797.546639] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Acquired lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.546639] env[65121]: DEBUG nova.network.neutron [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Refreshing network info cache for port ac5e737f-9b34-4d3a-896b-163829980cce {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 797.549075] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:cc:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15538852-1a3f-4f71-b4a9-4923c5837c4f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac5e737f-9b34-4d3a-896b-163829980cce', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:04:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '41278529-8bd2-44a1-97c8-03967faa3ff7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '183f83c8-107c-47b0-9081-d9a9173bc9b3', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.560650] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 797.560650] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.560650] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9408f68-c5b6-4aa8-b468-b1e072ef3899 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.588062] env[65121]: DEBUG oslo_vmware.api [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106417, 'name': PowerOffVM_Task, 'duration_secs': 0.249028} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.589054] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.589259] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.589348] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afa4c175-615f-4fff-a586-befacbe0d27c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.595037] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.595037] env[65121]: value = "task-5106419" [ 797.595037] env[65121]: _type = "Task" [ 797.595037] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.605446] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106419, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.817931] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 797.818391] env[65121]: DEBUG nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Instance network_info: |[{"id": "4304bec2-f452-4be6-aeda-3ad8354a4547", "address": "fa:16:3e:71:b3:4e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4304bec2-f4", "ovs_interfaceid": "4304bec2-f452-4be6-aeda-3ad8354a4547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 797.818954] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:b3:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4304bec2-f452-4be6-aeda-3ad8354a4547', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.828350] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating folder: Project (00383aa3355e438cb703c2b86c7917f2). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.832097] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40a89adf-6d22-4b81-b05a-e2763aa10841 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.914703] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.915555] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.916145] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Deleting the datastore file [datastore2] 00362477-c89e-4f60-98a4-d4928081d55e {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.918133] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b75ad24-2303-4e15-a5e9-6246562f824a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.920727] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created folder: Project (00383aa3355e438cb703c2b86c7917f2) in parent group-v993268. [ 797.920969] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating folder: Instances. Parent ref: group-v993411. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.924610] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98cbd841-4826-4961-9bdf-6c1d2fd60e2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.933088] env[65121]: DEBUG oslo_vmware.api [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for the task: (returnval){ [ 797.933088] env[65121]: value = "task-5106422" [ 797.933088] env[65121]: _type = "Task" [ 797.933088] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.939351] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created folder: Instances in parent group-v993411. [ 797.939735] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 797.951324] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.951848] env[65121]: DEBUG oslo_vmware.api [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.953182] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-145f3c5b-d460-40cd-aa80-ea4263ddb114 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.975435] env[65121]: DEBUG nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 797.989028] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.989028] env[65121]: value = "task-5106424" [ 797.989028] env[65121]: _type = "Task" [ 797.989028] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.997487] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106424, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.062030] env[65121]: WARNING neutronclient.v2_0.client [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.062759] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.062759] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.108079] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106419, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.125442] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7b71ab-fa29-483c-a891-c9285dccc9ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.135425] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3634779e-705a-4e38-be8c-b89a47744cb1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.176789] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fc09c2-9b17-4b40-8a1d-c71a297f7993 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.187044] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f71c39-6e80-433c-806e-c5a2f0e3bc76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.203029] env[65121]: DEBUG nova.compute.provider_tree [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.340950] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.341371] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.415588] env[65121]: WARNING neutronclient.v2_0.client [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.415588] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.415777] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.444135] env[65121]: DEBUG oslo_vmware.api [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Task: {'id': task-5106422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296011} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.444390] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 798.445682] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 798.445682] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 798.445682] env[65121]: INFO nova.compute.manager [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Took 1.43 seconds to destroy the instance on the hypervisor. [ 798.445682] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 798.445682] env[65121]: DEBUG nova.compute.manager [-] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 798.445682] env[65121]: DEBUG nova.network.neutron [-] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 798.445933] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.446282] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.446550] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.499289] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106424, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.506714] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.535791] env[65121]: DEBUG nova.network.neutron [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Updated VIF entry in instance network info cache for port ac5e737f-9b34-4d3a-896b-163829980cce. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 798.536277] env[65121]: DEBUG nova.network.neutron [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Updating instance_info_cache with network_info: [{"id": "ac5e737f-9b34-4d3a-896b-163829980cce", "address": "fa:16:3e:83:cc:41", "network": {"id": "b62e3e62-2227-4ef5-b7db-529e77bc8e38", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-28468943", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.222", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15538852-1a3f-4f71-b4a9-4923c5837c4f", "external-id": "nsx-vlan-transportzone-165", "segmentation_id": 165, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac5e737f-9b", "ovs_interfaceid": "ac5e737f-9b34-4d3a-896b-163829980cce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "address": "fa:16:3e:2a:04:d8", "network": {"id": "92b9c32c-2ead-47d9-9c56-0065346f5d96", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-698410639", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41278529-8bd2-44a1-97c8-03967faa3ff7", "external-id": "nsx-vlan-transportzone-749", "segmentation_id": 749, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap183f83c8-10", "ovs_interfaceid": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 798.538333] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.606562] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106419, 'name': CreateVM_Task, 'duration_secs': 0.598984} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.606755] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.607330] env[65121]: WARNING neutronclient.v2_0.client [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.607683] env[65121]: WARNING neutronclient.v2_0.client [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.608020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.608171] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.609182] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 798.609371] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d16f4deb-d4fe-440c-9310-11cca8502d5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.615687] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 798.615687] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dbf969-4fc1-9e75-5415-e9352203743d" [ 798.615687] env[65121]: _type = "Task" [ 798.615687] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.625261] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dbf969-4fc1-9e75-5415-e9352203743d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.749168] env[65121]: DEBUG nova.scheduler.client.report [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 798.749821] env[65121]: DEBUG nova.compute.provider_tree [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 77 to 78 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 798.749821] env[65121]: DEBUG nova.compute.provider_tree [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.941717] env[65121]: DEBUG nova.compute.manager [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Received event network-changed-cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 798.941892] env[65121]: DEBUG nova.compute.manager [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Refreshing instance network info cache due to event network-changed-cd096b57-4e55-46f3-b323-cc951f405db2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 798.942095] env[65121]: DEBUG oslo_concurrency.lockutils [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Acquiring lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.942268] env[65121]: DEBUG oslo_concurrency.lockutils [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Acquired lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.942339] env[65121]: DEBUG nova.network.neutron [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Refreshing network info cache for port cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 798.962906] env[65121]: INFO nova.compute.manager [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Rescuing [ 798.963181] env[65121]: DEBUG oslo_concurrency.lockutils [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.963329] env[65121]: DEBUG oslo_concurrency.lockutils [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.963398] env[65121]: DEBUG nova.network.neutron [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 799.001092] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106424, 'name': CreateVM_Task, 'duration_secs': 0.811393} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.001323] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.001982] env[65121]: WARNING neutronclient.v2_0.client [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.002690] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.041934] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Releasing lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.042273] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Received event network-vif-plugged-4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 799.042567] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Acquiring lock "4a72fedd-b114-468e-8f34-0caec6ce73fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.042799] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.042967] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.043168] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] No waiting events found dispatching network-vif-plugged-4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 799.043336] env[65121]: WARNING nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Received unexpected event network-vif-plugged-4304bec2-f452-4be6-aeda-3ad8354a4547 for instance with vm_state building and task_state spawning. [ 799.043783] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Received event network-changed-4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 799.044224] env[65121]: DEBUG nova.compute.manager [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Refreshing instance network info cache due to event network-changed-4304bec2-f452-4be6-aeda-3ad8354a4547. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 799.044224] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Acquiring lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.045053] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Acquired lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.045053] env[65121]: DEBUG nova.network.neutron [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Refreshing network info cache for port 4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 799.046259] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Acquired lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.046259] env[65121]: DEBUG nova.network.neutron [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Refreshing network info cache for port 183f83c8-107c-47b0-9081-d9a9173bc9b3 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 799.128121] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dbf969-4fc1-9e75-5415-e9352203743d, 'name': SearchDatastore_Task, 'duration_secs': 0.014797} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.129113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.129113] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.129113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.129113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.129319] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.129560] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.129908] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 799.130172] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da6c6524-bbe3-4c79-a720-c89d857d7f2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.132919] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52ef88bb-719c-4184-a60c-e83b5e411b04 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.140047] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 799.140047] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f319cb-7c7f-bfbc-f536-513d66cc22fa" [ 799.140047] env[65121]: _type = "Task" [ 799.140047] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.147461] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.147689] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.149792] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d50f941-7370-4850-ad75-c9eba8afa9a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.157306] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f319cb-7c7f-bfbc-f536-513d66cc22fa, 'name': SearchDatastore_Task, 'duration_secs': 0.012213} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.158371] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.158686] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.158950] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.162936] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 799.162936] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e7767c-6fb7-55f1-f0b2-e641f769b21f" [ 799.162936] env[65121]: _type = "Task" [ 799.162936] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.173445] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e7767c-6fb7-55f1-f0b2-e641f769b21f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.257180] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.305s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.257941] env[65121]: DEBUG nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 799.262552] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.327s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.265964] env[65121]: INFO nova.compute.claims [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.274656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "19b1f3f9-842e-4150-8890-b0b22393c3af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.274656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.274656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "19b1f3f9-842e-4150-8890-b0b22393c3af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.274656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.274656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.278199] env[65121]: DEBUG nova.network.neutron [-] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 799.280072] env[65121]: INFO nova.compute.manager [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Terminating instance [ 799.393997] env[65121]: DEBUG nova.compute.manager [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 799.393997] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6dc62b-775d-4a8f-8c32-1c6c4a11874f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.445347] env[65121]: WARNING neutronclient.v2_0.client [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.446030] env[65121]: WARNING openstack [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.446361] env[65121]: WARNING openstack [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.466180] env[65121]: WARNING neutronclient.v2_0.client [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.466868] env[65121]: WARNING openstack [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.467266] env[65121]: WARNING openstack [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.551162] env[65121]: WARNING neutronclient.v2_0.client [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.551162] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.551162] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.561072] env[65121]: WARNING neutronclient.v2_0.client [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.561971] env[65121]: WARNING openstack [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.562406] env[65121]: WARNING openstack [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.674091] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e7767c-6fb7-55f1-f0b2-e641f769b21f, 'name': SearchDatastore_Task, 'duration_secs': 0.011734} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.674954] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d2e0c93-3345-43b6-a3ad-7204a80a0c39 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.680865] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 799.680865] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52705e7e-e6ea-7573-4a58-98a28d9a5d1d" [ 799.680865] env[65121]: _type = "Task" [ 799.680865] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.689855] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52705e7e-e6ea-7573-4a58-98a28d9a5d1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.769271] env[65121]: DEBUG nova.compute.manager [req-6921f58f-0009-4f8d-a0fe-cfb5ec59f56a req-abf59289-ca18-4d13-b19b-bfcc773403f5 service nova] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Received event network-vif-deleted-7381b4f6-e691-437c-9b26-59e3192fe6ed {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 799.770857] env[65121]: DEBUG nova.compute.utils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 799.775363] env[65121]: DEBUG nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 799.775712] env[65121]: DEBUG nova.network.neutron [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 799.776204] env[65121]: WARNING neutronclient.v2_0.client [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.776663] env[65121]: WARNING neutronclient.v2_0.client [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.777623] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.778266] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.789600] env[65121]: INFO nova.compute.manager [-] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Took 1.34 seconds to deallocate network for instance. [ 799.791861] env[65121]: DEBUG nova.compute.manager [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 799.791861] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 799.797471] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df717f41-278e-4fb8-8172-7495e7df6a62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.812065] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 799.812422] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0c13677-0e79-46f9-90e4-de70e91c205c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.820880] env[65121]: DEBUG oslo_vmware.api [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 799.820880] env[65121]: value = "task-5106426" [ 799.820880] env[65121]: _type = "Task" [ 799.820880] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.831572] env[65121]: DEBUG oslo_vmware.api [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106426, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.911206] env[65121]: INFO nova.compute.manager [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] instance snapshotting [ 799.911206] env[65121]: DEBUG nova.objects.instance [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'flavor' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.103244] env[65121]: WARNING openstack [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.103918] env[65121]: WARNING openstack [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.111844] env[65121]: WARNING openstack [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.115336] env[65121]: WARNING openstack [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.153366] env[65121]: DEBUG nova.policy [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '978d227edd2649a48d366fea8c8a567e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2794dfddd4814fdc8a1a2e9eae0d8b51', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 800.193074] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52705e7e-e6ea-7573-4a58-98a28d9a5d1d, 'name': SearchDatastore_Task, 'duration_secs': 0.01421} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.193426] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 800.193692] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c1632a96-f334-4978-9f31-97f9329577e5/c1632a96-f334-4978-9f31-97f9329577e5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.193957] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 800.194149] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.194419] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-721c5ee6-bc9b-47ce-b5f6-71365c2942fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.197495] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b99fafde-f85e-44c8-96f3-1ac1267f6266 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.205458] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 800.205458] env[65121]: value = "task-5106427" [ 800.205458] env[65121]: _type = "Task" [ 800.205458] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.210724] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 800.210884] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 800.212017] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ddd4521-5531-4e4b-8108-6c5f9f4b88be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.217725] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106427, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.221024] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 800.221024] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5218d8bc-1213-ce2c-40e4-70c12d98990b" [ 800.221024] env[65121]: _type = "Task" [ 800.221024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.230095] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5218d8bc-1213-ce2c-40e4-70c12d98990b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.233714] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.234105] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.249693] env[65121]: WARNING openstack [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.250101] env[65121]: WARNING openstack [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.275854] env[65121]: DEBUG nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 800.310367] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 800.335481] env[65121]: DEBUG oslo_vmware.api [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106426, 'name': PowerOffVM_Task, 'duration_secs': 0.224541} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.336480] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 800.336649] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 800.341882] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bec8027-2d39-4829-97c1-d2bb631d0e01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.424210] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfd32f3-d3a4-4cbd-aeb4-09c881d3dc2d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.450298] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 800.450802] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 800.450802] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Deleting the datastore file [datastore1] 19b1f3f9-842e-4150-8890-b0b22393c3af {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 800.455071] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8a4c87-c0b2-4739-85d7-34e3ceef0c3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.458860] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-effe96a3-f018-4571-9b60-5baccc117d12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.478516] env[65121]: DEBUG oslo_vmware.api [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 800.478516] env[65121]: value = "task-5106429" [ 800.478516] env[65121]: _type = "Task" [ 800.478516] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.490830] env[65121]: DEBUG oslo_vmware.api [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106429, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.717847] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106427, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.737414] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5218d8bc-1213-ce2c-40e4-70c12d98990b, 'name': SearchDatastore_Task, 'duration_secs': 0.010445} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.738494] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c84fa4d-41b0-4225-9915-d9f60610dc52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.749429] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 800.749429] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d634ac-e9ee-6727-fd0b-e36e55da4721" [ 800.749429] env[65121]: _type = "Task" [ 800.749429] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.763791] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d634ac-e9ee-6727-fd0b-e36e55da4721, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.811543] env[65121]: DEBUG nova.network.neutron [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Successfully created port: 203dc15c-4972-4c3d-9e50-af138f9134fe {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 800.885799] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.886048] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 800.958185] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23da450-70aa-4bad-90fb-d9e0eb06cf0d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.967368] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b873f35b-5a69-41d9-a0cb-785cd936ddbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.980469] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 800.981951] env[65121]: WARNING neutronclient.v2_0.client [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.982590] env[65121]: WARNING openstack [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.982985] env[65121]: WARNING openstack [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.991426] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f44eb23c-75a9-4c54-a553-b7b13efe9181 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.024561] env[65121]: WARNING neutronclient.v2_0.client [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.025513] env[65121]: WARNING openstack [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.025893] env[65121]: WARNING openstack [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.035549] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b8cb57-72a9-4ae6-9d76-c7357ad6d127 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.042885] env[65121]: DEBUG oslo_vmware.api [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106429, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.376161} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.044906] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.045666] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 801.045666] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 801.045666] env[65121]: INFO nova.compute.manager [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Took 1.25 seconds to destroy the instance on the hypervisor. [ 801.045797] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 801.046077] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 801.046077] env[65121]: value = "task-5106430" [ 801.046077] env[65121]: _type = "Task" [ 801.046077] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.046427] env[65121]: DEBUG nova.compute.manager [-] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 801.046427] env[65121]: DEBUG nova.network.neutron [-] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 801.046675] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.047319] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.047747] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.063862] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e50849b-7282-4706-81d3-842dc212ecbd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.073693] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106430, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.084982] env[65121]: DEBUG nova.compute.provider_tree [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.217461] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106427, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566385} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.217745] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c1632a96-f334-4978-9f31-97f9329577e5/c1632a96-f334-4978-9f31-97f9329577e5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.218015] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.218206] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bf063f0-e2b4-4a45-951a-fb1ccc54414e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.226387] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 801.226387] env[65121]: value = "task-5106431" [ 801.226387] env[65121]: _type = "Task" [ 801.226387] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.238044] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106431, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.261425] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d634ac-e9ee-6727-fd0b-e36e55da4721, 'name': SearchDatastore_Task, 'duration_secs': 0.012417} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.262993] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 801.262993] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 4a72fedd-b114-468e-8f34-0caec6ce73fa/4a72fedd-b114-468e-8f34-0caec6ce73fa.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 801.262993] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27142a3b-8a8e-461b-a81e-f77c579a6ad9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.271024] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 801.271024] env[65121]: value = "task-5106432" [ 801.271024] env[65121]: _type = "Task" [ 801.271024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.281962] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.289993] env[65121]: DEBUG nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 801.335419] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 801.337846] env[65121]: DEBUG nova.virt.hardware [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 801.340671] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9626e654-012a-4487-aec7-327ad3f1695e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.350973] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfd85b7-edb0-4942-828e-e4e23bae964b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.401585] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.401585] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.401585] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.401585] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.401585] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.401585] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.401585] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 801.401585] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 801.567846] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106430, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.588549] env[65121]: DEBUG nova.scheduler.client.report [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 801.739321] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106431, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142432} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.739778] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.740684] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0f968b-9dde-4e33-b20d-fb3665edca50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.776569] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] c1632a96-f334-4978-9f31-97f9329577e5/c1632a96-f334-4978-9f31-97f9329577e5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.777123] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e224347a-eecd-4065-9318-1ad18177e297 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.805595] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106432, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.807794] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 801.807794] env[65121]: value = "task-5106434" [ 801.807794] env[65121]: _type = "Task" [ 801.807794] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.817966] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106434, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.906151] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.000133] env[65121]: WARNING neutronclient.v2_0.client [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.000891] env[65121]: WARNING openstack [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.001300] env[65121]: WARNING openstack [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.029201] env[65121]: WARNING neutronclient.v2_0.client [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.029763] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.030124] env[65121]: WARNING openstack [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.067732] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106430, 'name': CreateSnapshot_Task, 'duration_secs': 0.646909} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.072615] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 802.073900] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c304e90-2778-4a37-a12d-2a785b7a2240 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.096744] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.833s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.096744] env[65121]: DEBUG nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 802.099054] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.672s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.099316] env[65121]: DEBUG nova.objects.instance [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lazy-loading 'resources' on Instance uuid 3f69040e-7df3-4535-a3a7-90f3348ef346 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.186486] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.287049] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106432, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574911} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.287319] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 4a72fedd-b114-468e-8f34-0caec6ce73fa/4a72fedd-b114-468e-8f34-0caec6ce73fa.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 802.287523] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.287771] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21e583d7-56f0-422c-bb44-44298029348b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.295270] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 802.295270] env[65121]: value = "task-5106435" [ 802.295270] env[65121]: _type = "Task" [ 802.295270] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.305889] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106435, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.321237] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106434, 'name': ReconfigVM_Task, 'duration_secs': 0.397436} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.321582] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Reconfigured VM instance instance-00000031 to attach disk [datastore2] c1632a96-f334-4978-9f31-97f9329577e5/c1632a96-f334-4978-9f31-97f9329577e5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.322367] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cde573f-ac01-4dad-9bb7-62150c90acac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.330155] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 802.330155] env[65121]: value = "task-5106436" [ 802.330155] env[65121]: _type = "Task" [ 802.330155] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.339976] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106436, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.598225] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 802.598786] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-44897b27-80b3-46a7-87eb-a7ed02aab548 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.610500] env[65121]: DEBUG nova.network.neutron [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Updating instance_info_cache with network_info: [{"id": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "address": "fa:16:3e:81:57:23", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1740aa3b-24", "ovs_interfaceid": "1740aa3b-2428-4bc3-8888-c4edc0734c60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.615020] env[65121]: DEBUG nova.compute.utils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 802.621174] env[65121]: DEBUG nova.network.neutron [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Successfully updated port: 203dc15c-4972-4c3d-9e50-af138f9134fe {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 802.621460] env[65121]: DEBUG nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 802.621963] env[65121]: DEBUG nova.network.neutron [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 802.621963] env[65121]: WARNING neutronclient.v2_0.client [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.622680] env[65121]: WARNING neutronclient.v2_0.client [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 802.623358] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.623758] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.641205] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 802.641205] env[65121]: value = "task-5106437" [ 802.641205] env[65121]: _type = "Task" [ 802.641205] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.652202] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106437, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.654736] env[65121]: DEBUG nova.network.neutron [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updated VIF entry in instance network info cache for port cd096b57-4e55-46f3-b323-cc951f405db2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 802.654991] env[65121]: DEBUG nova.network.neutron [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updating instance_info_cache with network_info: [{"id": "cd096b57-4e55-46f3-b323-cc951f405db2", "address": "fa:16:3e:40:1f:eb", "network": {"id": "7106221e-7e9a-4351-aaa6-776cba11b989", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1587920485-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce90017874734825b9fbdce06eb12a00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd096b57-4e", "ovs_interfaceid": "cd096b57-4e55-46f3-b323-cc951f405db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.806245] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106435, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079064} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.809041] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.810310] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff7b86c-2fe6-47eb-ac52-5a270d8f88b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.837774] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 4a72fedd-b114-468e-8f34-0caec6ce73fa/4a72fedd-b114-468e-8f34-0caec6ce73fa.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.840881] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd76c9e3-ba5e-4faf-95a0-1095ee2656c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.871477] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106436, 'name': Rename_Task, 'duration_secs': 0.161816} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.875462] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.875462] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 802.875462] env[65121]: value = "task-5106438" [ 802.875462] env[65121]: _type = "Task" [ 802.875462] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.875462] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eac3794f-814e-41d2-a9f4-30ea20cdf3b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.888915] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 802.888915] env[65121]: value = "task-5106439" [ 802.888915] env[65121]: _type = "Task" [ 802.888915] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.890053] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106438, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.902622] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.904937] env[65121]: DEBUG nova.network.neutron [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Updated VIF entry in instance network info cache for port 183f83c8-107c-47b0-9081-d9a9173bc9b3. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 802.905410] env[65121]: DEBUG nova.network.neutron [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Updating instance_info_cache with network_info: [{"id": "ac5e737f-9b34-4d3a-896b-163829980cce", "address": "fa:16:3e:83:cc:41", "network": {"id": "b62e3e62-2227-4ef5-b7db-529e77bc8e38", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-28468943", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.222", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15538852-1a3f-4f71-b4a9-4923c5837c4f", "external-id": "nsx-vlan-transportzone-165", "segmentation_id": 165, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac5e737f-9b", "ovs_interfaceid": "ac5e737f-9b34-4d3a-896b-163829980cce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "address": "fa:16:3e:2a:04:d8", "network": {"id": "92b9c32c-2ead-47d9-9c56-0065346f5d96", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-698410639", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "bc44d2096bff489193193973fe3a3550", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41278529-8bd2-44a1-97c8-03967faa3ff7", "external-id": "nsx-vlan-transportzone-749", "segmentation_id": 749, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap183f83c8-10", "ovs_interfaceid": "183f83c8-107c-47b0-9081-d9a9173bc9b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.959249] env[65121]: DEBUG nova.network.neutron [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updated VIF entry in instance network info cache for port 4304bec2-f452-4be6-aeda-3ad8354a4547. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 802.959632] env[65121]: DEBUG nova.network.neutron [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updating instance_info_cache with network_info: [{"id": "4304bec2-f452-4be6-aeda-3ad8354a4547", "address": "fa:16:3e:71:b3:4e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4304bec2-f4", "ovs_interfaceid": "4304bec2-f452-4be6-aeda-3ad8354a4547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.964197] env[65121]: DEBUG nova.policy [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bf232407a7346c98ea64ed673a179f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b35af4e2b6844a9081f34466661c83b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 803.119839] env[65121]: DEBUG nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 803.124853] env[65121]: DEBUG oslo_concurrency.lockutils [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "refresh_cache-ab41ae41-e69d-47fb-a31a-16748fd3b0f2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.136970] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.137214] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquired lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 803.137399] env[65121]: DEBUG nova.network.neutron [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 803.156630] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106437, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.160285] env[65121]: DEBUG oslo_concurrency.lockutils [req-af45b305-7b48-457c-8175-063957d7e1c6 req-b473f4f5-38ff-4804-993c-93bfcf0715ef service nova] Releasing lock "refresh_cache-19b1f3f9-842e-4150-8890-b0b22393c3af" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.366019] env[65121]: DEBUG nova.network.neutron [-] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 803.379045] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260581e7-d8c7-42cf-b7b2-bac5a0c049df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.404143] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106438, 'name': ReconfigVM_Task, 'duration_secs': 0.38017} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.404143] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9686f01a-70ea-47b1-9de2-952b9a6f97bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.408024] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 4a72fedd-b114-468e-8f34-0caec6ce73fa/4a72fedd-b114-468e-8f34-0caec6ce73fa.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.412503] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7c7db04-da32-428b-8653-ea2d904905e3 req-edbef235-fc69-470b-8c27-6535b4d474cf service nova] Releasing lock "refresh_cache-c1632a96-f334-4978-9f31-97f9329577e5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.412887] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c082ba3-a6a1-4c21-8788-a1e110c492bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.418798] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106439, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.453203] env[65121]: DEBUG nova.network.neutron [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Successfully created port: 9127a7ec-601e-4377-adfb-134eaaa9c000 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 803.456933] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d1a4eb-4f8d-4bfe-8770-6a431fb7247a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.460268] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 803.460268] env[65121]: value = "task-5106440" [ 803.460268] env[65121]: _type = "Task" [ 803.460268] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.470499] env[65121]: DEBUG oslo_concurrency.lockutils [req-63b5a684-8f88-4665-a701-50d1b796ea42 req-b579e49d-69ce-4d55-9d90-7ac7f41e792c service nova] Releasing lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.472576] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89345a8b-d654-4f90-a4d0-35c55abc2a1a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.483178] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106440, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.497784] env[65121]: DEBUG nova.compute.provider_tree [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.641429] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 803.641847] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 803.659887] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106437, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.870912] env[65121]: INFO nova.compute.manager [-] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Took 2.82 seconds to deallocate network for instance. [ 803.881138] env[65121]: DEBUG nova.network.neutron [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 803.905177] env[65121]: DEBUG oslo_vmware.api [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106439, 'name': PowerOnVM_Task, 'duration_secs': 0.609709} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.905177] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.905177] env[65121]: INFO nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Took 12.01 seconds to spawn the instance on the hypervisor. [ 803.905177] env[65121]: DEBUG nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 803.905703] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd3dc47-0d2c-4f20-80bd-c2d9fd810c5a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.973875] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106440, 'name': Rename_Task, 'duration_secs': 0.180651} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.974714] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.975097] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22efb5f6-b406-4e48-9778-7e758540d505 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.984672] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 803.984672] env[65121]: value = "task-5106442" [ 803.984672] env[65121]: _type = "Task" [ 803.984672] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.996018] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106442, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.001199] env[65121]: DEBUG nova.scheduler.client.report [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 804.134068] env[65121]: DEBUG nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 804.161825] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106437, 'name': CloneVM_Task, 'duration_secs': 1.490322} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.164161] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Created linked-clone VM from snapshot [ 804.165315] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2017a2eb-a9bc-465e-80e7-9cc1caacfe84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.170944] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 804.171213] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 804.171421] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 804.171531] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 804.171662] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 804.171802] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 804.171998] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.172161] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 804.172320] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 804.172472] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 804.172633] env[65121]: DEBUG nova.virt.hardware [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 804.176958] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eca92db-cbc9-4617-a822-292d2fe120d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.183688] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Uploading image 1caf99bb-23c5-4c78-a9b4-8ea36bb253c1 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 804.193834] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a83bbb-efa2-4d43-be4b-197ce97e18c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.221585] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 804.221585] env[65121]: value = "vm-993415" [ 804.221585] env[65121]: _type = "VirtualMachine" [ 804.221585] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 804.221915] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0b947f23-8bb7-4588-8a42-568e788185b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.229936] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease: (returnval){ [ 804.229936] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521e330c-27e1-0383-663f-9a32c5acd859" [ 804.229936] env[65121]: _type = "HttpNfcLease" [ 804.229936] env[65121]: } obtained for exporting VM: (result){ [ 804.229936] env[65121]: value = "vm-993415" [ 804.229936] env[65121]: _type = "VirtualMachine" [ 804.229936] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 804.230305] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the lease: (returnval){ [ 804.230305] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521e330c-27e1-0383-663f-9a32c5acd859" [ 804.230305] env[65121]: _type = "HttpNfcLease" [ 804.230305] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 804.238823] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 804.238823] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521e330c-27e1-0383-663f-9a32c5acd859" [ 804.238823] env[65121]: _type = "HttpNfcLease" [ 804.238823] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 804.388532] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 804.432352] env[65121]: INFO nova.compute.manager [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Took 43.97 seconds to build instance. [ 804.497145] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106442, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.508789] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.410s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.511491] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.203s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.513269] env[65121]: INFO nova.compute.claims [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.538698] env[65121]: INFO nova.scheduler.client.report [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Deleted allocations for instance 3f69040e-7df3-4535-a3a7-90f3348ef346 [ 804.677327] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.677722] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28d9e040-d86b-4177-969b-d87764aa693a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.688433] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 804.688433] env[65121]: value = "task-5106444" [ 804.688433] env[65121]: _type = "Task" [ 804.688433] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.705027] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106444, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.742330] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 804.742330] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521e330c-27e1-0383-663f-9a32c5acd859" [ 804.742330] env[65121]: _type = "HttpNfcLease" [ 804.742330] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 804.743759] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 804.743759] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521e330c-27e1-0383-663f-9a32c5acd859" [ 804.743759] env[65121]: _type = "HttpNfcLease" [ 804.743759] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 804.743759] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd5718b-2a4b-4254-bd0e-16cb6771cf13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.754771] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a466-9a66-e5c7-e9fe-5035b7eb7a97/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 804.755073] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a466-9a66-e5c7-e9fe-5035b7eb7a97/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 804.875771] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6105a54c-190c-4d46-95fe-ad5697db5291 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.933812] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6048366d-4f1a-4b8f-85bd-7ada0ebf6de0 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "c1632a96-f334-4978-9f31-97f9329577e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.834s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.991708] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.992132] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.004791] env[65121]: DEBUG oslo_vmware.api [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106442, 'name': PowerOnVM_Task, 'duration_secs': 0.685792} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.005217] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.005603] env[65121]: INFO nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Took 10.25 seconds to spawn the instance on the hypervisor. [ 805.005769] env[65121]: DEBUG nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 805.006866] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b99201-90fe-4af9-ac34-1a6c3a314b3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.050362] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1cb200c1-b8f8-4371-abc9-167faae6bcc0 tempest-TenantUsagesTestJSON-459545200 tempest-TenantUsagesTestJSON-459545200-project-member] Lock "3f69040e-7df3-4535-a3a7-90f3348ef346" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.678s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.200803] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106444, 'name': PowerOffVM_Task, 'duration_secs': 0.223651} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.202279] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.202279] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93b1c6f-84a0-4546-acdc-54cbb8b68496 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.206254] env[65121]: WARNING neutronclient.v2_0.client [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.207056] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.207443] env[65121]: WARNING openstack [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.239508] env[65121]: DEBUG nova.network.neutron [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Successfully updated port: 9127a7ec-601e-4377-adfb-134eaaa9c000 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 805.241926] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26aa474c-024b-452e-b577-40283830271c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.289506] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.289978] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5633f4a2-c0ae-4781-93fd-7b5984e992a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.300282] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 805.300282] env[65121]: value = "task-5106445" [ 805.300282] env[65121]: _type = "Task" [ 805.300282] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.312179] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 805.312584] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.312863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.313101] env[65121]: DEBUG oslo_concurrency.lockutils [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.313367] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.313753] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32a6abc4-ab27-465a-a93e-ba8112930974 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.334061] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.334061] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.335449] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87dea9f4-e94a-4625-9ee6-0bafa23bf15d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.342697] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 805.342697] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ef887d-6914-fed5-5520-7f574fc4a75c" [ 805.342697] env[65121]: _type = "Task" [ 805.342697] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.352994] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ef887d-6914-fed5-5520-7f574fc4a75c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.368687] env[65121]: DEBUG nova.network.neutron [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Updating instance_info_cache with network_info: [{"id": "203dc15c-4972-4c3d-9e50-af138f9134fe", "address": "fa:16:3e:b5:be:bb", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203dc15c-49", "ovs_interfaceid": "203dc15c-4972-4c3d-9e50-af138f9134fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.439493] env[65121]: DEBUG nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 805.536020] env[65121]: INFO nova.compute.manager [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Took 43.95 seconds to build instance. [ 805.747428] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "refresh_cache-90cfd88c-286e-4c87-bf27-909873359997" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.747428] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "refresh_cache-90cfd88c-286e-4c87-bf27-909873359997" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.747428] env[65121]: DEBUG nova.network.neutron [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 805.838294] env[65121]: DEBUG nova.compute.manager [req-95f8c84c-6603-421d-84ed-70a1baba1663 req-301d2d58-a751-4958-835c-eca9245897a3 service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Received event network-vif-plugged-203dc15c-4972-4c3d-9e50-af138f9134fe {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 805.838654] env[65121]: DEBUG oslo_concurrency.lockutils [req-95f8c84c-6603-421d-84ed-70a1baba1663 req-301d2d58-a751-4958-835c-eca9245897a3 service nova] Acquiring lock "f694fcb6-053b-4649-ac63-7fa98b1373eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.839021] env[65121]: DEBUG oslo_concurrency.lockutils [req-95f8c84c-6603-421d-84ed-70a1baba1663 req-301d2d58-a751-4958-835c-eca9245897a3 service nova] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.839767] env[65121]: DEBUG oslo_concurrency.lockutils [req-95f8c84c-6603-421d-84ed-70a1baba1663 req-301d2d58-a751-4958-835c-eca9245897a3 service nova] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.839767] env[65121]: DEBUG nova.compute.manager [req-95f8c84c-6603-421d-84ed-70a1baba1663 req-301d2d58-a751-4958-835c-eca9245897a3 service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] No waiting events found dispatching network-vif-plugged-203dc15c-4972-4c3d-9e50-af138f9134fe {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 805.839767] env[65121]: WARNING nova.compute.manager [req-95f8c84c-6603-421d-84ed-70a1baba1663 req-301d2d58-a751-4958-835c-eca9245897a3 service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Received unexpected event network-vif-plugged-203dc15c-4972-4c3d-9e50-af138f9134fe for instance with vm_state building and task_state spawning. [ 805.860306] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ef887d-6914-fed5-5520-7f574fc4a75c, 'name': SearchDatastore_Task, 'duration_secs': 0.029412} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.860306] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df6e98e6-480f-4458-835e-b343f10d03b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.867759] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 805.867759] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524f9ab2-3733-74d2-c9c7-65e9cf1b90ce" [ 805.867759] env[65121]: _type = "Task" [ 805.867759] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.876206] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Releasing lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.876206] env[65121]: DEBUG nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Instance network_info: |[{"id": "203dc15c-4972-4c3d-9e50-af138f9134fe", "address": "fa:16:3e:b5:be:bb", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203dc15c-49", "ovs_interfaceid": "203dc15c-4972-4c3d-9e50-af138f9134fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 805.876445] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:be:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '203dc15c-4972-4c3d-9e50-af138f9134fe', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.885502] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Creating folder: Project (2794dfddd4814fdc8a1a2e9eae0d8b51). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.886863] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-707243e4-3ac2-47c5-8e71-311bfe997188 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.893120] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524f9ab2-3733-74d2-c9c7-65e9cf1b90ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.908396] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Created folder: Project (2794dfddd4814fdc8a1a2e9eae0d8b51) in parent group-v993268. [ 805.908622] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Creating folder: Instances. Parent ref: group-v993417. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.908890] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f0c2b17-9efd-4dcd-93e2-3399cab6cb1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.922439] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Created folder: Instances in parent group-v993417. [ 805.922808] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 805.923022] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.923249] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-897d16ca-42a5-4297-bef4-4f0645cfe6a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.955317] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.955317] env[65121]: value = "task-5106449" [ 805.955317] env[65121]: _type = "Task" [ 805.955317] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.965300] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106449, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.969474] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.039793] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e7b16afc-3d7b-4cf2-bc4f-256653aabc1a tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.162s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.252861] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.253379] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.283168] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8261a294-9860-4a05-90fe-407bd7736ad4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.293485] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f71eec9-da00-4064-9cd2-fc8515b5f9d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.332258] env[65121]: DEBUG nova.network.neutron [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 806.336349] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa82052-134e-48cd-8361-477f6491bcb7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.345859] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe5771f-ce1a-4d53-a8d3-34e1318d0bdb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.361955] env[65121]: DEBUG nova.compute.provider_tree [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.385848] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524f9ab2-3733-74d2-c9c7-65e9cf1b90ce, 'name': SearchDatastore_Task, 'duration_secs': 0.029768} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.387588] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.388114] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.395334] env[65121]: DEBUG oslo_concurrency.lockutils [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.395700] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. {{(pid=65121) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 806.396730] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7bb7cf03-4fd3-46f3-a225-dcc72ea65bb4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.407624] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 806.407624] env[65121]: value = "task-5106450" [ 806.407624] env[65121]: _type = "Task" [ 806.407624] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.420147] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106450, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.474149] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106449, 'name': CreateVM_Task, 'duration_secs': 0.371164} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.474474] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.475074] env[65121]: WARNING neutronclient.v2_0.client [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 806.478109] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.479087] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.479087] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 806.479087] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1094fa69-8fc4-4436-87c6-1ffba58680a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.486287] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 806.486287] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520c5a85-c1a8-5060-dbf8-e9c39f3ec2b0" [ 806.486287] env[65121]: _type = "Task" [ 806.486287] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.498663] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520c5a85-c1a8-5060-dbf8-e9c39f3ec2b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.503940] env[65121]: WARNING neutronclient.v2_0.client [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 806.503940] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.504204] env[65121]: WARNING openstack [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.546436] env[65121]: DEBUG nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 806.557701] env[65121]: DEBUG nova.compute.manager [req-c4f066d6-c93d-45ca-8fd6-d353aae2d2fe req-a32aebc3-3b96-4253-a102-9c93f765bd4b service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Received event network-vif-plugged-9127a7ec-601e-4377-adfb-134eaaa9c000 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 806.558421] env[65121]: DEBUG oslo_concurrency.lockutils [req-c4f066d6-c93d-45ca-8fd6-d353aae2d2fe req-a32aebc3-3b96-4253-a102-9c93f765bd4b service nova] Acquiring lock "90cfd88c-286e-4c87-bf27-909873359997-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.558735] env[65121]: DEBUG oslo_concurrency.lockutils [req-c4f066d6-c93d-45ca-8fd6-d353aae2d2fe req-a32aebc3-3b96-4253-a102-9c93f765bd4b service nova] Lock "90cfd88c-286e-4c87-bf27-909873359997-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.559627] env[65121]: DEBUG oslo_concurrency.lockutils [req-c4f066d6-c93d-45ca-8fd6-d353aae2d2fe req-a32aebc3-3b96-4253-a102-9c93f765bd4b service nova] Lock "90cfd88c-286e-4c87-bf27-909873359997-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.559834] env[65121]: DEBUG nova.compute.manager [req-c4f066d6-c93d-45ca-8fd6-d353aae2d2fe req-a32aebc3-3b96-4253-a102-9c93f765bd4b service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] No waiting events found dispatching network-vif-plugged-9127a7ec-601e-4377-adfb-134eaaa9c000 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 806.560041] env[65121]: WARNING nova.compute.manager [req-c4f066d6-c93d-45ca-8fd6-d353aae2d2fe req-a32aebc3-3b96-4253-a102-9c93f765bd4b service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Received unexpected event network-vif-plugged-9127a7ec-601e-4377-adfb-134eaaa9c000 for instance with vm_state building and task_state spawning. [ 806.649913] env[65121]: DEBUG nova.network.neutron [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Updating instance_info_cache with network_info: [{"id": "9127a7ec-601e-4377-adfb-134eaaa9c000", "address": "fa:16:3e:f6:de:f9", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9127a7ec-60", "ovs_interfaceid": "9127a7ec-601e-4377-adfb-134eaaa9c000", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 806.866062] env[65121]: DEBUG nova.scheduler.client.report [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 806.922616] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106450, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.939289] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "93721179-407a-43d9-b0bf-157433a09519" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.939289] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "93721179-407a-43d9-b0bf-157433a09519" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.999233] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520c5a85-c1a8-5060-dbf8-e9c39f3ec2b0, 'name': SearchDatastore_Task, 'duration_secs': 0.021799} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.000347] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.000347] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.000347] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.000347] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.000786] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.000893] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5a400c5-4c80-4724-a406-d2c72cc86c6c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.011497] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.013234] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.014249] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7a45176-6a13-4067-a04f-964a0ef981ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.024511] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 807.024511] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52eedabc-82a4-dd74-e874-338367a06850" [ 807.024511] env[65121]: _type = "Task" [ 807.024511] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.035051] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52eedabc-82a4-dd74-e874-338367a06850, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.075225] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.154023] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "refresh_cache-90cfd88c-286e-4c87-bf27-909873359997" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.154531] env[65121]: DEBUG nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Instance network_info: |[{"id": "9127a7ec-601e-4377-adfb-134eaaa9c000", "address": "fa:16:3e:f6:de:f9", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9127a7ec-60", "ovs_interfaceid": "9127a7ec-601e-4377-adfb-134eaaa9c000", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 807.155225] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:de:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9127a7ec-601e-4377-adfb-134eaaa9c000', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.163952] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 807.164822] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.165206] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12644521-e186-40a0-914a-5bc38d734c7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.190704] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.190704] env[65121]: value = "task-5106451" [ 807.190704] env[65121]: _type = "Task" [ 807.190704] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.200573] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106451, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.372841] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.861s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 807.374229] env[65121]: DEBUG nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 807.378117] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.211s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 807.380620] env[65121]: DEBUG nova.objects.instance [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lazy-loading 'resources' on Instance uuid 50083ee5-9655-4cab-9d50-04a97baac626 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.420846] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106450, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628319} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.421375] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. [ 807.423101] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0771781-774a-434d-9754-e9a2db365d2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.456396] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.457396] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2de0815-ac1d-4143-9805-c718376996e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.483848] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 807.483848] env[65121]: value = "task-5106452" [ 807.483848] env[65121]: _type = "Task" [ 807.483848] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.494779] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106452, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.539535] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52eedabc-82a4-dd74-e874-338367a06850, 'name': SearchDatastore_Task, 'duration_secs': 0.017158} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.539535] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-635b7d37-b3ec-4669-9ceb-e3b9f4588001 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.546788] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 807.546788] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5246be7a-64db-ce37-899e-174ad00a8833" [ 807.546788] env[65121]: _type = "Task" [ 807.546788] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.558537] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246be7a-64db-ce37-899e-174ad00a8833, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.701217] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106451, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.886371] env[65121]: DEBUG nova.compute.utils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 807.892281] env[65121]: DEBUG nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 807.892584] env[65121]: DEBUG nova.network.neutron [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 807.896225] env[65121]: WARNING neutronclient.v2_0.client [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.896225] env[65121]: WARNING neutronclient.v2_0.client [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 807.896225] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.896999] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.988638] env[65121]: DEBUG nova.policy [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d4b5db5a0c44a3fbdebf4bdd6d21145', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9209f0dd88bc43129ae91f27d8212239', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 808.004723] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106452, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.069497] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246be7a-64db-ce37-899e-174ad00a8833, 'name': SearchDatastore_Task, 'duration_secs': 0.015366} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.069852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.071427] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] f694fcb6-053b-4649-ac63-7fa98b1373eb/f694fcb6-053b-4649-ac63-7fa98b1373eb.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.072255] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ada05c87-1f8e-4ea0-aeab-72178ebe564b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.085024] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 808.085024] env[65121]: value = "task-5106454" [ 808.085024] env[65121]: _type = "Task" [ 808.085024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.106269] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106454, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.209418] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106451, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.408711] env[65121]: DEBUG nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 808.502733] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106452, 'name': ReconfigVM_Task, 'duration_secs': 0.529143} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.503065] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Reconfigured VM instance instance-00000030 to attach disk [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.503949] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5eaaee-402a-4b2f-9800-ccd90a9c89d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.541709] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4788957e-17f9-45bb-98d5-8e7e3cad180e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.555186] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "c1632a96-f334-4978-9f31-97f9329577e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 808.555496] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "c1632a96-f334-4978-9f31-97f9329577e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.555729] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "c1632a96-f334-4978-9f31-97f9329577e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 808.555961] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "c1632a96-f334-4978-9f31-97f9329577e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.556164] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "c1632a96-f334-4978-9f31-97f9329577e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.566379] env[65121]: INFO nova.compute.manager [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Terminating instance [ 808.575634] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 808.575634] env[65121]: value = "task-5106455" [ 808.575634] env[65121]: _type = "Task" [ 808.575634] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.589865] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106455, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.608334] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106454, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.687593] env[65121]: DEBUG nova.network.neutron [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Successfully created port: ab950b26-0d72-483b-87f5-4686e5bd3c88 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 808.707601] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106451, 'name': CreateVM_Task, 'duration_secs': 1.511095} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.707858] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.708509] env[65121]: WARNING neutronclient.v2_0.client [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.708948] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.709127] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.709489] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 808.709797] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ebbbf0e-2692-4f3c-b84c-d6e74ff6b35b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.716237] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 808.716237] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c17c65-04b9-8765-a621-ffd6553582fd" [ 808.716237] env[65121]: _type = "Task" [ 808.716237] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.730309] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c17c65-04b9-8765-a621-ffd6553582fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.804010] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d36fb1f-40d8-4ca3-a2d7-de041ac25960 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.813705] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07477d7f-05e4-4f49-9aab-caa03a0627db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.847370] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e806cd7f-a2e0-4c9c-91c6-15e4338e7228 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.857825] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f149deee-3f79-4990-8304-69119d91f359 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.877926] env[65121]: DEBUG nova.compute.provider_tree [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.067684] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.067990] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.068668] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.068923] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.069134] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.075809] env[65121]: DEBUG nova.compute.manager [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 809.076223] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.076891] env[65121]: INFO nova.compute.manager [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Terminating instance [ 809.079812] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d033e279-e35c-4a8c-a3ef-95b3d3ae994a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.099786] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.099986] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106455, 'name': ReconfigVM_Task, 'duration_secs': 0.2841} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.101613] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb709854-909a-4259-ac55-731c21719dbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.103491] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.104848] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbdbb9f8-92bc-4106-b451-4d8c588784ea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.112246] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106454, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.118946] env[65121]: DEBUG oslo_vmware.api [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 809.118946] env[65121]: value = "task-5106456" [ 809.118946] env[65121]: _type = "Task" [ 809.118946] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.120719] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 809.120719] env[65121]: value = "task-5106457" [ 809.120719] env[65121]: _type = "Task" [ 809.120719] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.137051] env[65121]: DEBUG oslo_vmware.api [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106456, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.144743] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.237265] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c17c65-04b9-8765-a621-ffd6553582fd, 'name': SearchDatastore_Task, 'duration_secs': 0.017387} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.237755] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.238077] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.238894] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.239113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.240062] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.242044] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae188d91-9520-47c5-9643-3fafc0342f72 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.254861] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.255244] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.256320] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-226a1660-a769-4c0b-ac2f-0e23e7c8e697 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.264809] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 809.264809] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b4d1e8-5039-ca0c-d484-4250b087b2c6" [ 809.264809] env[65121]: _type = "Task" [ 809.264809] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.279913] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b4d1e8-5039-ca0c-d484-4250b087b2c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.380325] env[65121]: DEBUG nova.scheduler.client.report [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 809.422364] env[65121]: DEBUG nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 809.453529] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 809.453801] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 809.453955] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 809.454238] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 809.454437] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 809.454625] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 809.454886] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.455122] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 809.455338] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 809.455537] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 809.455740] env[65121]: DEBUG nova.virt.hardware [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 809.456833] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cadd9afe-f6d9-4284-8767-2fc982ab2e68 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.467065] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55dd253-dc45-41c4-94fe-da8f08c2121d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.600665] env[65121]: DEBUG nova.compute.manager [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 809.601110] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.601281] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106454, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.206846} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.602157] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b4ebfa-f45b-4fbe-a839-a03e1166c683 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.609180] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] f694fcb6-053b-4649-ac63-7fa98b1373eb/f694fcb6-053b-4649-ac63-7fa98b1373eb.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.609574] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.609982] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a384fcba-be4f-4c79-b60c-725630e9a140 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.618668] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.620701] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e36b41a-af48-4680-bce5-3883786d353f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.622303] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 809.622303] env[65121]: value = "task-5106458" [ 809.622303] env[65121]: _type = "Task" [ 809.622303] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.637129] env[65121]: DEBUG oslo_vmware.api [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 809.637129] env[65121]: value = "task-5106459" [ 809.637129] env[65121]: _type = "Task" [ 809.637129] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.654946] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106457, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.655242] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.655451] env[65121]: DEBUG oslo_vmware.api [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106456, 'name': PowerOffVM_Task, 'duration_secs': 0.311679} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.656173] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.656345] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.656606] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c86ff3f9-6922-4ea0-8b36-2b6662cb8e3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.661623] env[65121]: DEBUG oslo_vmware.api [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.780333] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b4d1e8-5039-ca0c-d484-4250b087b2c6, 'name': SearchDatastore_Task, 'duration_secs': 0.016654} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.782133] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 809.782458] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 809.782458] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Deleting the datastore file [datastore2] c1632a96-f334-4978-9f31-97f9329577e5 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.782758] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-508355f2-6ede-4c6a-aa5d-9ecb8953d497 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.785379] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd8c59b6-17b9-4af2-bd8d-5c526832358d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.795140] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 809.795140] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f10ae6-1049-aec3-17e9-e0afea0c3bb6" [ 809.795140] env[65121]: _type = "Task" [ 809.795140] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.796654] env[65121]: DEBUG oslo_vmware.api [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for the task: (returnval){ [ 809.796654] env[65121]: value = "task-5106461" [ 809.796654] env[65121]: _type = "Task" [ 809.796654] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.808897] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f10ae6-1049-aec3-17e9-e0afea0c3bb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.811994] env[65121]: DEBUG oslo_vmware.api [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.887348] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.509s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.890089] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.810s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.890350] env[65121]: DEBUG nova.objects.instance [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lazy-loading 'resources' on Instance uuid 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.923478] env[65121]: INFO nova.scheduler.client.report [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Deleted allocations for instance 50083ee5-9655-4cab-9d50-04a97baac626 [ 810.133023] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116534} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.136485] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.137641] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fca32d4-084e-4aa2-a046-f7754900a9bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.147615] env[65121]: DEBUG oslo_vmware.api [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106457, 'name': PowerOnVM_Task, 'duration_secs': 0.694925} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.157190] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.169601] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] f694fcb6-053b-4649-ac63-7fa98b1373eb/f694fcb6-053b-4649-ac63-7fa98b1373eb.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.170872] env[65121]: DEBUG nova.compute.manager [None req-211f6707-3388-44c8-929f-acb9b67a3327 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 810.174267] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-935f0c52-ffc5-4f5c-a0d8-7aff725ab64d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.189608] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890c27a2-51cf-4849-8685-a4f62688c3a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.192707] env[65121]: DEBUG oslo_vmware.api [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106459, 'name': PowerOffVM_Task, 'duration_secs': 0.253412} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.192975] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.193151] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.195430] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3389108-81b0-4898-9de6-478c68f230a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.202356] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 810.202356] env[65121]: value = "task-5106463" [ 810.202356] env[65121]: _type = "Task" [ 810.202356] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.215994] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106463, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.277513] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.277763] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.277947] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Deleting the datastore file [datastore2] cd86f815-4b46-45b1-9a1c-ab2e28f63626 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.278262] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de980cbe-c2e2-4f19-93c9-73fab3393b6a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.286538] env[65121]: DEBUG oslo_vmware.api [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 810.286538] env[65121]: value = "task-5106465" [ 810.286538] env[65121]: _type = "Task" [ 810.286538] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.296278] env[65121]: DEBUG oslo_vmware.api [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106465, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.311905] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f10ae6-1049-aec3-17e9-e0afea0c3bb6, 'name': SearchDatastore_Task, 'duration_secs': 0.017475} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.315967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.316302] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 90cfd88c-286e-4c87-bf27-909873359997/90cfd88c-286e-4c87-bf27-909873359997.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.318054] env[65121]: DEBUG oslo_vmware.api [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Task: {'id': task-5106461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284418} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.318054] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d55420e-4451-4730-ae7a-451becb73ebf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.319909] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.320211] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.320418] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.320592] env[65121]: INFO nova.compute.manager [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Took 1.24 seconds to destroy the instance on the hypervisor. [ 810.320911] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 810.321252] env[65121]: DEBUG nova.compute.manager [-] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 810.321367] env[65121]: DEBUG nova.network.neutron [-] [instance: c1632a96-f334-4978-9f31-97f9329577e5] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 810.321686] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.322329] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.322649] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.338197] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 810.338197] env[65121]: value = "task-5106466" [ 810.338197] env[65121]: _type = "Task" [ 810.338197] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.348474] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106466, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.431830] env[65121]: DEBUG nova.compute.manager [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Received event network-changed-203dc15c-4972-4c3d-9e50-af138f9134fe {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 810.432103] env[65121]: DEBUG nova.compute.manager [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Refreshing instance network info cache due to event network-changed-203dc15c-4972-4c3d-9e50-af138f9134fe. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 810.432386] env[65121]: DEBUG oslo_concurrency.lockutils [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Acquiring lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.432598] env[65121]: DEBUG oslo_concurrency.lockutils [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Acquired lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.432826] env[65121]: DEBUG nova.network.neutron [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Refreshing network info cache for port 203dc15c-4972-4c3d-9e50-af138f9134fe {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 810.443967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df5ad2be-5abc-4d7a-8a9d-742ae9c39058 tempest-VolumesAdminNegativeTest-147738353 tempest-VolumesAdminNegativeTest-147738353-project-member] Lock "50083ee5-9655-4cab-9d50-04a97baac626" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.253s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.531442] env[65121]: DEBUG nova.network.neutron [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Successfully updated port: ab950b26-0d72-483b-87f5-4686e5bd3c88 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 810.634122] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.732162] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106463, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.806165] env[65121]: DEBUG oslo_vmware.api [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106465, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265014} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.806165] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.806645] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.806645] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.806645] env[65121]: INFO nova.compute.manager [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Took 1.21 seconds to destroy the instance on the hypervisor. [ 810.807147] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 810.807147] env[65121]: DEBUG nova.compute.manager [-] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 810.807306] env[65121]: DEBUG nova.network.neutron [-] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 810.807461] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.808037] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.808291] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.850559] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106466, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.943037] env[65121]: WARNING neutronclient.v2_0.client [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.943893] env[65121]: WARNING openstack [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.944469] env[65121]: WARNING openstack [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.036728] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.036875] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquired lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.037073] env[65121]: DEBUG nova.network.neutron [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 811.054736] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4902a258-fb60-4332-9e6b-96a6dc12a3fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.065335] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f155e3-502c-4c92-b4b9-a917f296535e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.102745] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e931429-7c29-4c5f-933d-cae5d5e9b708 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.112461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5faeedd5-eb57-426a-871a-4da6c46b1148 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.128940] env[65121]: DEBUG nova.compute.provider_tree [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.219845] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106463, 'name': ReconfigVM_Task, 'duration_secs': 0.585221} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.220183] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Reconfigured VM instance instance-00000033 to attach disk [datastore2] f694fcb6-053b-4649-ac63-7fa98b1373eb/f694fcb6-053b-4649-ac63-7fa98b1373eb.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.220977] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f1942bf-2859-4433-b8e7-978b19ba6ab6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.229618] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 811.229618] env[65121]: value = "task-5106467" [ 811.229618] env[65121]: _type = "Task" [ 811.229618] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.239804] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106467, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.354861] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106466, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.683735} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.355214] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 90cfd88c-286e-4c87-bf27-909873359997/90cfd88c-286e-4c87-bf27-909873359997.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.355476] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.355751] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8fb8bdc-5090-4cea-9504-f932bffde8a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.367036] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 811.367036] env[65121]: value = "task-5106468" [ 811.367036] env[65121]: _type = "Task" [ 811.367036] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.378164] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.541459] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.541459] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.633684] env[65121]: DEBUG nova.scheduler.client.report [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 811.741439] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106467, 'name': Rename_Task, 'duration_secs': 0.256053} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.742832] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.742832] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7ff62b8-2248-4ecd-ab9e-9913fd77eef0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.751256] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 811.751256] env[65121]: value = "task-5106469" [ 811.751256] env[65121]: _type = "Task" [ 811.751256] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.764945] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.827492] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 811.881563] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074248} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.881842] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.882730] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7996a220-155d-482e-bdd5-18c99bc93cf4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.889097] env[65121]: DEBUG nova.network.neutron [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 811.919226] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 90cfd88c-286e-4c87-bf27-909873359997/90cfd88c-286e-4c87-bf27-909873359997.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.920131] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78745486-a47e-480b-83bb-9cc6ceda0d4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.946378] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 811.946378] env[65121]: value = "task-5106470" [ 811.946378] env[65121]: _type = "Task" [ 811.946378] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.961214] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106470, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.141353] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.249s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.142782] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.492s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.145367] env[65121]: INFO nova.compute.claims [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.186961] env[65121]: INFO nova.scheduler.client.report [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted allocations for instance 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f [ 812.265740] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106469, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.421160] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.421614] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.458028] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106470, 'name': ReconfigVM_Task, 'duration_secs': 0.490338} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.458338] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 90cfd88c-286e-4c87-bf27-909873359997/90cfd88c-286e-4c87-bf27-909873359997.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.458980] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17c30fa8-95b9-4fed-90c9-fe5d63289960 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.470980] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 812.470980] env[65121]: value = "task-5106471" [ 812.470980] env[65121]: _type = "Task" [ 812.470980] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.479830] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106471, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.525562] env[65121]: WARNING openstack [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.526068] env[65121]: WARNING openstack [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.642028] env[65121]: DEBUG nova.network.neutron [-] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 812.704692] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0a8c359-b620-46d5-bc42-bb5441b69912 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.938s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.712768] env[65121]: WARNING neutronclient.v2_0.client [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.713608] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.713971] env[65121]: WARNING openstack [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.762787] env[65121]: DEBUG oslo_vmware.api [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106469, 'name': PowerOnVM_Task, 'duration_secs': 0.641093} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.763167] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.763463] env[65121]: INFO nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Took 11.47 seconds to spawn the instance on the hypervisor. [ 812.763664] env[65121]: DEBUG nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 812.764726] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1399203d-0841-410c-8d1a-d870c49198e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.817234] env[65121]: WARNING neutronclient.v2_0.client [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.818619] env[65121]: WARNING openstack [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.818619] env[65121]: WARNING openstack [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.981397] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106471, 'name': Rename_Task, 'duration_secs': 0.339913} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.981518] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.981775] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c608ae57-4f52-4d82-b8d2-66364ddda575 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.983920] env[65121]: DEBUG nova.network.neutron [-] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 812.992473] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 812.992473] env[65121]: value = "task-5106472" [ 812.992473] env[65121]: _type = "Task" [ 812.992473] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.013123] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106472, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.017437] env[65121]: DEBUG nova.network.neutron [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Updating instance_info_cache with network_info: [{"id": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "address": "fa:16:3e:af:f0:fd", "network": {"id": "f16ea9d5-3016-4746-8747-20bc1a0d5a8e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-801656862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9209f0dd88bc43129ae91f27d8212239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab950b26-0d", "ovs_interfaceid": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 813.115545] env[65121]: DEBUG nova.network.neutron [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Updated VIF entry in instance network info cache for port 203dc15c-4972-4c3d-9e50-af138f9134fe. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 813.115900] env[65121]: DEBUG nova.network.neutron [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Updating instance_info_cache with network_info: [{"id": "203dc15c-4972-4c3d-9e50-af138f9134fe", "address": "fa:16:3e:b5:be:bb", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203dc15c-49", "ovs_interfaceid": "203dc15c-4972-4c3d-9e50-af138f9134fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 813.143032] env[65121]: INFO nova.compute.manager [-] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Took 2.82 seconds to deallocate network for instance. [ 813.174045] env[65121]: DEBUG nova.compute.manager [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Received event network-changed-9127a7ec-601e-4377-adfb-134eaaa9c000 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 813.177483] env[65121]: DEBUG nova.compute.manager [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Refreshing instance network info cache due to event network-changed-9127a7ec-601e-4377-adfb-134eaaa9c000. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 813.177613] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Acquiring lock "refresh_cache-90cfd88c-286e-4c87-bf27-909873359997" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.180033] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Acquired lock "refresh_cache-90cfd88c-286e-4c87-bf27-909873359997" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 813.180222] env[65121]: DEBUG nova.network.neutron [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Refreshing network info cache for port 9127a7ec-601e-4377-adfb-134eaaa9c000 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 813.289712] env[65121]: INFO nova.compute.manager [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Took 42.44 seconds to build instance. [ 813.490105] env[65121]: INFO nova.compute.manager [-] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Took 2.68 seconds to deallocate network for instance. [ 813.509145] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106472, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.519908] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Releasing lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 813.520280] env[65121]: DEBUG nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Instance network_info: |[{"id": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "address": "fa:16:3e:af:f0:fd", "network": {"id": "f16ea9d5-3016-4746-8747-20bc1a0d5a8e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-801656862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9209f0dd88bc43129ae91f27d8212239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab950b26-0d", "ovs_interfaceid": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 813.521366] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:f0:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ab950b26-0d72-483b-87f5-4686e5bd3c88', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.529829] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Creating folder: Project (9209f0dd88bc43129ae91f27d8212239). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.530155] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9b960fa-8a9b-49d1-9488-507f35808330 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.546857] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Created folder: Project (9209f0dd88bc43129ae91f27d8212239) in parent group-v993268. [ 813.547133] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Creating folder: Instances. Parent ref: group-v993421. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.549393] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cde4f762-4936-4c29-a0c9-7826fe72b31c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.566266] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Created folder: Instances in parent group-v993421. [ 813.566485] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 813.566713] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 813.566942] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35c4b095-c2fd-4eb5-b737-189406bc4b1a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.591876] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.591876] env[65121]: value = "task-5106475" [ 813.591876] env[65121]: _type = "Task" [ 813.591876] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.602786] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106475, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.623800] env[65121]: DEBUG oslo_concurrency.lockutils [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] Releasing lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 813.623800] env[65121]: DEBUG nova.compute.manager [req-8bb8c99c-c9fc-4a5e-ba88-b43ceac08c36 req-c35330a3-3909-4064-a341-5a91c670f43c service nova] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Received event network-vif-deleted-cd096b57-4e55-46f3-b323-cc951f405db2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 813.659982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.685997] env[65121]: WARNING neutronclient.v2_0.client [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 813.686641] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.687317] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.798046] env[65121]: DEBUG oslo_concurrency.lockutils [None req-68b29d36-23fe-4289-b4fa-416425881d90 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.903s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.861421] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92a7f7e-0866-487d-9062-16deb3ea0b77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.873924] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4251ee1c-798e-411d-9fa3-503b46cdff79 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.919825] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9e0b10-29e5-4d90-a542-eb2bb571f30b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.926029] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.926029] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.939491] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3deffc5b-feef-4a48-9b7b-0a90fc4e7418 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.960893] env[65121]: DEBUG nova.compute.provider_tree [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.003436] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.013877] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106472, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.071189] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "7fcab9cb-9a0f-4a50-9244-687889443a31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.071189] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.110423] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106475, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.150155] env[65121]: WARNING neutronclient.v2_0.client [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.150914] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.151579] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.306305] env[65121]: DEBUG nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 814.341041] env[65121]: DEBUG nova.network.neutron [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Updated VIF entry in instance network info cache for port 9127a7ec-601e-4377-adfb-134eaaa9c000. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 814.341261] env[65121]: DEBUG nova.network.neutron [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Updating instance_info_cache with network_info: [{"id": "9127a7ec-601e-4377-adfb-134eaaa9c000", "address": "fa:16:3e:f6:de:f9", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9127a7ec-60", "ovs_interfaceid": "9127a7ec-601e-4377-adfb-134eaaa9c000", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 814.362389] env[65121]: DEBUG nova.compute.manager [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Received event network-vif-plugged-ab950b26-0d72-483b-87f5-4686e5bd3c88 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 814.362546] env[65121]: DEBUG oslo_concurrency.lockutils [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Acquiring lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.362756] env[65121]: DEBUG oslo_concurrency.lockutils [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.362984] env[65121]: DEBUG oslo_concurrency.lockutils [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.363145] env[65121]: DEBUG nova.compute.manager [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] No waiting events found dispatching network-vif-plugged-ab950b26-0d72-483b-87f5-4686e5bd3c88 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 814.364295] env[65121]: WARNING nova.compute.manager [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Received unexpected event network-vif-plugged-ab950b26-0d72-483b-87f5-4686e5bd3c88 for instance with vm_state building and task_state spawning. [ 814.364295] env[65121]: DEBUG nova.compute.manager [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Received event network-changed-ab950b26-0d72-483b-87f5-4686e5bd3c88 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 814.364560] env[65121]: DEBUG nova.compute.manager [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Refreshing instance network info cache due to event network-changed-ab950b26-0d72-483b-87f5-4686e5bd3c88. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 814.364677] env[65121]: DEBUG oslo_concurrency.lockutils [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Acquiring lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.368153] env[65121]: DEBUG oslo_concurrency.lockutils [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Acquired lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.369182] env[65121]: DEBUG nova.network.neutron [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Refreshing network info cache for port ab950b26-0d72-483b-87f5-4686e5bd3c88 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 814.465203] env[65121]: DEBUG nova.scheduler.client.report [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 814.513457] env[65121]: DEBUG oslo_vmware.api [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106472, 'name': PowerOnVM_Task, 'duration_secs': 1.061964} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.513724] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.513956] env[65121]: INFO nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Took 10.38 seconds to spawn the instance on the hypervisor. [ 814.514241] env[65121]: DEBUG nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 814.515340] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b8fafb-a854-4be7-8f84-5466993b4027 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.603511] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106475, 'name': CreateVM_Task, 'duration_secs': 0.549982} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.604262] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 814.604968] env[65121]: WARNING neutronclient.v2_0.client [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.606508] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.609876] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.609876] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 814.609876] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-995f4db3-33df-4ebd-a6c5-3fff4a15f31b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.616109] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 814.616109] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520646aa-fae1-fe74-baed-7b5462e1c488" [ 814.616109] env[65121]: _type = "Task" [ 814.616109] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.627029] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520646aa-fae1-fe74-baed-7b5462e1c488, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.822885] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.823230] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.823434] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.823603] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.823751] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.827707] env[65121]: INFO nova.compute.manager [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Terminating instance [ 814.846239] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.847501] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Releasing lock "refresh_cache-90cfd88c-286e-4c87-bf27-909873359997" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 814.847762] env[65121]: DEBUG nova.compute.manager [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Received event network-changed-4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 814.847876] env[65121]: DEBUG nova.compute.manager [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Refreshing instance network info cache due to event network-changed-4304bec2-f452-4be6-aeda-3ad8354a4547. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 814.848165] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Acquiring lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.848257] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Acquired lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.848355] env[65121]: DEBUG nova.network.neutron [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Refreshing network info cache for port 4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 814.872034] env[65121]: WARNING neutronclient.v2_0.client [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.872767] env[65121]: WARNING openstack [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.873200] env[65121]: WARNING openstack [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.971089] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.828s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.971184] env[65121]: DEBUG nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 814.974348] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.291s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.978622] env[65121]: INFO nova.compute.claims [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.044321] env[65121]: INFO nova.compute.manager [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Took 43.14 seconds to build instance. [ 815.096276] env[65121]: WARNING openstack [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.096589] env[65121]: WARNING openstack [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.116318] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a466-9a66-e5c7-e9fe-5035b7eb7a97/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 815.116318] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9423d2d-6573-49dc-ac3d-9549895b6e8f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.127184] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a466-9a66-e5c7-e9fe-5035b7eb7a97/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 815.127636] env[65121]: ERROR oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a466-9a66-e5c7-e9fe-5035b7eb7a97/disk-0.vmdk due to incomplete transfer. [ 815.131525] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ed439ab2-f933-4b1d-b2af-2ed302eaec1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.133528] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520646aa-fae1-fe74-baed-7b5462e1c488, 'name': SearchDatastore_Task, 'duration_secs': 0.014865} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.134101] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 815.134435] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.135078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.135361] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 815.135637] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.136428] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86be14d3-6d75-4892-86f6-786650e7eb88 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.145021] env[65121]: DEBUG oslo_vmware.rw_handles [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e9a466-9a66-e5c7-e9fe-5035b7eb7a97/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 815.145021] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Uploaded image 1caf99bb-23c5-4c78-a9b4-8ea36bb253c1 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 815.147408] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 815.147945] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-449bcb8e-2a67-44ec-8b18-482b285ea339 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.158078] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 815.158078] env[65121]: value = "task-5106476" [ 815.158078] env[65121]: _type = "Task" [ 815.158078] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.158667] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.158986] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.167025] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31b3e704-61b8-40d8-92d6-55546338fa3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.172305] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 815.172305] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524cd760-fcf2-6f2b-869e-588bbda07135" [ 815.172305] env[65121]: _type = "Task" [ 815.172305] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.177296] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106476, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.190258] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524cd760-fcf2-6f2b-869e-588bbda07135, 'name': SearchDatastore_Task, 'duration_secs': 0.012655} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.191130] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7273ce61-eee3-482c-b2bf-977110703003 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.198691] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 815.198691] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5233d76d-9eef-36fb-ceb9-cfa29b5f9622" [ 815.198691] env[65121]: _type = "Task" [ 815.198691] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.214206] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5233d76d-9eef-36fb-ceb9-cfa29b5f9622, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.271853] env[65121]: WARNING neutronclient.v2_0.client [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.272591] env[65121]: WARNING openstack [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.273575] env[65121]: WARNING openstack [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.337027] env[65121]: DEBUG nova.compute.manager [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 815.337027] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.337027] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc32bb3b-0c9c-4ea6-88f6-49e8baa6ba2d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.346133] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.346277] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c320cbdb-a480-450e-93d4-3e555af727b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.351375] env[65121]: WARNING neutronclient.v2_0.client [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.352385] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.352786] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.362466] env[65121]: DEBUG oslo_vmware.api [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 815.362466] env[65121]: value = "task-5106477" [ 815.362466] env[65121]: _type = "Task" [ 815.362466] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.374049] env[65121]: DEBUG oslo_vmware.api [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106477, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.491669] env[65121]: DEBUG nova.compute.utils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 815.494402] env[65121]: DEBUG nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 815.494402] env[65121]: DEBUG nova.network.neutron [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 815.494402] env[65121]: WARNING neutronclient.v2_0.client [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.494402] env[65121]: WARNING neutronclient.v2_0.client [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.494834] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.495281] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.532225] env[65121]: DEBUG nova.network.neutron [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Updated VIF entry in instance network info cache for port ab950b26-0d72-483b-87f5-4686e5bd3c88. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 815.532590] env[65121]: DEBUG nova.network.neutron [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Updating instance_info_cache with network_info: [{"id": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "address": "fa:16:3e:af:f0:fd", "network": {"id": "f16ea9d5-3016-4746-8747-20bc1a0d5a8e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-801656862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9209f0dd88bc43129ae91f27d8212239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab950b26-0d", "ovs_interfaceid": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 815.554078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e5fa95f-94b5-4e0c-b0bf-be09ec4bda91 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "90cfd88c-286e-4c87-bf27-909873359997" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.287s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.662108] env[65121]: DEBUG nova.policy [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1c51fc43b7f44ce9b023d8cb769e029', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89110618b6ef4ce68b2ca53e7e168139', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 815.683019] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106476, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.710510] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5233d76d-9eef-36fb-ceb9-cfa29b5f9622, 'name': SearchDatastore_Task, 'duration_secs': 0.020173} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.710851] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 815.711218] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7/ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 815.711638] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7cd7cb6a-a809-4dda-b7fd-60201e2648cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.719838] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.722828] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.730094] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 815.730094] env[65121]: value = "task-5106478" [ 815.730094] env[65121]: _type = "Task" [ 815.730094] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.740476] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.810794] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.811106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.876192] env[65121]: DEBUG oslo_vmware.api [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106477, 'name': PowerOffVM_Task, 'duration_secs': 0.379242} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.876934] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.877194] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.877312] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43b89348-a0b8-4e76-a3ab-7fb750a35372 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.974693] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.974693] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.974693] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Deleting the datastore file [datastore1] 41af3ee5-0ee6-4957-bad0-5aef51d55bd5 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.974693] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57b69319-e58f-4178-bc49-39d460005de2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.982735] env[65121]: DEBUG oslo_vmware.api [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for the task: (returnval){ [ 815.982735] env[65121]: value = "task-5106480" [ 815.982735] env[65121]: _type = "Task" [ 815.982735] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.004119] env[65121]: DEBUG nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 816.007550] env[65121]: DEBUG oslo_vmware.api [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.035846] env[65121]: DEBUG oslo_concurrency.lockutils [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] Releasing lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 816.036176] env[65121]: DEBUG nova.compute.manager [req-66a8f255-6e09-4b70-a967-21ae643d9b1d req-fc02564d-3e22-4c57-bfea-1dbabd1727ed service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received event network-vif-deleted-183f83c8-107c-47b0-9081-d9a9173bc9b3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 816.058418] env[65121]: DEBUG nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 816.119559] env[65121]: DEBUG nova.compute.manager [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Received event network-changed {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 816.119812] env[65121]: DEBUG nova.compute.manager [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Refreshing instance network info cache due to event network-changed. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 816.121982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Acquiring lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.121982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Acquired lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.121982] env[65121]: DEBUG nova.network.neutron [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 816.181332] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106476, 'name': Destroy_Task, 'duration_secs': 0.779412} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.185410] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Destroyed the VM [ 816.185723] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 816.187271] env[65121]: DEBUG nova.network.neutron [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Successfully created port: 6e8b4013-341c-4478-b8a7-1b7c43349ad8 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 816.189822] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-82f4c5b1-3f4f-4331-a16d-a9f4db74fced {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.200806] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 816.200806] env[65121]: value = "task-5106481" [ 816.200806] env[65121]: _type = "Task" [ 816.200806] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.217222] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.239404] env[65121]: WARNING neutronclient.v2_0.client [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.240283] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.241636] env[65121]: WARNING openstack [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.263712] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106478, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.503356] env[65121]: DEBUG oslo_vmware.api [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Task: {'id': task-5106480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.473627} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.503688] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.503909] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.504120] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.504302] env[65121]: INFO nova.compute.manager [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Took 1.17 seconds to destroy the instance on the hypervisor. [ 816.504629] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 816.504860] env[65121]: DEBUG nova.compute.manager [-] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 816.505193] env[65121]: DEBUG nova.network.neutron [-] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 816.505375] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.506022] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.506359] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.530300] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Acquiring lock "1b6ad893-e013-4fd5-a829-535d1ea6e001" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.530570] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.583764] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.624040] env[65121]: WARNING neutronclient.v2_0.client [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.624802] env[65121]: WARNING openstack [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.625498] env[65121]: WARNING openstack [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.718267] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.743830] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106478, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.776197} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.744111] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7/ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 816.744321] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.744574] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c42904f1-14cb-4e7e-b714-36d21998c4a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.751463] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcdc0d3-914d-4ee5-ba5f-05ab6a24f8c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.755619] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 816.755619] env[65121]: value = "task-5106482" [ 816.755619] env[65121]: _type = "Task" [ 816.755619] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.762997] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a300057b-86b3-429b-85cf-7a227962a466 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.770330] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.804554] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09512db3-c4da-4392-9f35-a5e500b59375 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.814016] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d3bce7-0a17-4793-a76c-6c97f7e2d970 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.835662] env[65121]: DEBUG nova.compute.provider_tree [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.031067] env[65121]: DEBUG nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 817.042131] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.062155] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 817.062509] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 817.062667] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 817.062911] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 817.063162] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 817.063335] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 817.063596] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.063697] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 817.063914] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 817.064143] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 817.064491] env[65121]: DEBUG nova.virt.hardware [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 817.065686] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5243e59d-af2e-4f2c-954b-9be861de75fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.076982] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65effe69-3a75-4144-b4d3-0d41f4e0c412 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.197135] env[65121]: DEBUG nova.network.neutron [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updated VIF entry in instance network info cache for port 4304bec2-f452-4be6-aeda-3ad8354a4547. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 817.197263] env[65121]: DEBUG nova.network.neutron [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updating instance_info_cache with network_info: [{"id": "4304bec2-f452-4be6-aeda-3ad8354a4547", "address": "fa:16:3e:71:b3:4e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4304bec2-f4", "ovs_interfaceid": "4304bec2-f452-4be6-aeda-3ad8354a4547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 817.212688] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106481, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.266674] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072961} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.266809] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.267825] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e31540-a491-45da-8f80-4bad2ef197b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.292861] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7/ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.293259] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc0628ba-ec5b-4682-804b-cb2ee78dea1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.316825] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 817.316825] env[65121]: value = "task-5106483" [ 817.316825] env[65121]: _type = "Task" [ 817.316825] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.326281] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.338552] env[65121]: DEBUG nova.scheduler.client.report [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 817.700019] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d04d41-72f9-4eb3-8dc4-7f49a1506634 req-c4b6a399-470b-45c5-9942-cf5f26a1d773 service nova] Releasing lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 817.714155] env[65121]: DEBUG oslo_vmware.api [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106481, 'name': RemoveSnapshot_Task, 'duration_secs': 1.206374} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.714466] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 817.714713] env[65121]: INFO nova.compute.manager [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Took 17.29 seconds to snapshot the instance on the hypervisor. [ 817.824498] env[65121]: WARNING openstack [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.825219] env[65121]: WARNING openstack [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.834990] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106483, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.845088] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.871s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.845621] env[65121]: DEBUG nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 817.848841] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.667s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.849080] env[65121]: DEBUG nova.objects.instance [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lazy-loading 'resources' on Instance uuid e2d32a5e-c350-4b2b-9243-c3b412193a82 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.024056] env[65121]: WARNING neutronclient.v2_0.client [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.024832] env[65121]: WARNING openstack [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.025369] env[65121]: WARNING openstack [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.179137] env[65121]: DEBUG nova.network.neutron [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Successfully updated port: 6e8b4013-341c-4478-b8a7-1b7c43349ad8 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 818.191246] env[65121]: DEBUG nova.network.neutron [-] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 818.266250] env[65121]: DEBUG nova.network.neutron [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Updating instance_info_cache with network_info: [{"id": "203dc15c-4972-4c3d-9e50-af138f9134fe", "address": "fa:16:3e:b5:be:bb", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203dc15c-49", "ovs_interfaceid": "203dc15c-4972-4c3d-9e50-af138f9134fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 818.278591] env[65121]: DEBUG nova.compute.manager [None req-a7105d7d-5a82-44b7-bd09-f466b2227bc1 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Found 1 images (rotation: 2) {{(pid=65121) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 818.331049] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106483, 'name': ReconfigVM_Task, 'duration_secs': 0.924499} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.331049] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Reconfigured VM instance instance-00000036 to attach disk [datastore2] ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7/ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.331315] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2bafa61f-8815-4b63-87fa-80e1cad29668 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.339964] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 818.339964] env[65121]: value = "task-5106484" [ 818.339964] env[65121]: _type = "Task" [ 818.339964] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.351053] env[65121]: DEBUG nova.compute.utils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 818.352447] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106484, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.352988] env[65121]: DEBUG nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 818.353662] env[65121]: DEBUG nova.network.neutron [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 818.353662] env[65121]: WARNING neutronclient.v2_0.client [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.353847] env[65121]: WARNING neutronclient.v2_0.client [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.354847] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.354847] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.439653] env[65121]: DEBUG nova.policy [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e773a3343a448f5acd3f08005e8ed45', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '40aee06073b04c25bc7b38fbd09856e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 818.685070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "refresh_cache-75bf0b85-db28-4efa-a517-d7256873b09d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.685614] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "refresh_cache-75bf0b85-db28-4efa-a517-d7256873b09d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 818.685951] env[65121]: DEBUG nova.network.neutron [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 818.696329] env[65121]: INFO nova.compute.manager [-] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Took 2.19 seconds to deallocate network for instance. [ 818.773126] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e9d96034-aad1-4100-9343-caeb9d2b3916 tempest-ServerExternalEventsTest-1371280343 tempest-ServerExternalEventsTest-1371280343-project] Releasing lock "refresh_cache-f694fcb6-053b-4649-ac63-7fa98b1373eb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.857527] env[65121]: DEBUG nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 818.859568] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106484, 'name': Rename_Task, 'duration_secs': 0.323805} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.859895] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 818.860511] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10738b84-c735-4acc-8d3f-0884b5c422b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.871874] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 818.871874] env[65121]: value = "task-5106485" [ 818.871874] env[65121]: _type = "Task" [ 818.871874] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.872642] env[65121]: DEBUG nova.network.neutron [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Successfully created port: 30bb9dd1-838c-45e3-80f2-580a96b705ff {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 818.887389] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106485, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.980118] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60088ad7-0b96-4d3a-b293-78bae932a116 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.989474] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4099ce84-60d9-4ecb-8219-c820bc906e92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.024145] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f779595-3983-4a01-8ede-15c5ba85123c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.033143] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8df0e2-1725-48fc-864b-f441c2831e49 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.049283] env[65121]: DEBUG nova.compute.provider_tree [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.103914] env[65121]: DEBUG nova.compute.manager [req-c74e393e-39b5-4664-8648-826ab57d3456 req-ba2fbc05-ada8-48d7-a571-85a8b9c08e6b service nova] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Received event network-vif-deleted-ac5e737f-9b34-4d3a-896b-163829980cce {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 819.104289] env[65121]: DEBUG nova.compute.manager [req-c74e393e-39b5-4664-8648-826ab57d3456 req-ba2fbc05-ada8-48d7-a571-85a8b9c08e6b service nova] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Received event network-vif-deleted-a56899a4-844b-4ea3-b942-aa09fda98ec3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 819.189603] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.190230] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.202270] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 819.253056] env[65121]: DEBUG nova.network.neutron [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 819.376044] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.376044] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.395668] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106485, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.506338] env[65121]: WARNING neutronclient.v2_0.client [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.506338] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.506732] env[65121]: WARNING openstack [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.552992] env[65121]: DEBUG nova.scheduler.client.report [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 819.635168] env[65121]: DEBUG nova.network.neutron [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Updating instance_info_cache with network_info: [{"id": "6e8b4013-341c-4478-b8a7-1b7c43349ad8", "address": "fa:16:3e:98:fc:60", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e8b4013-34", "ovs_interfaceid": "6e8b4013-341c-4478-b8a7-1b7c43349ad8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.868216] env[65121]: DEBUG nova.compute.manager [req-ead43b41-b823-45ea-ab29-818e32679c75 req-0f5bd6fa-62ed-4b64-b165-699507a29447 service nova] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Received event network-vif-deleted-68b68226-9044-4753-b2b7-cbae3b99512e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 819.875096] env[65121]: DEBUG nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 819.896273] env[65121]: DEBUG oslo_vmware.api [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106485, 'name': PowerOnVM_Task, 'duration_secs': 0.710982} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.896603] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 819.897774] env[65121]: INFO nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Took 10.47 seconds to spawn the instance on the hypervisor. [ 819.897774] env[65121]: DEBUG nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 819.898293] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477fafd1-36f6-42e0-bb95-bdd2b12b6f01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.919054] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 819.919666] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 819.919666] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 819.919817] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 819.919927] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 819.920143] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 819.920382] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.920531] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 819.920691] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 819.920950] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 819.921119] env[65121]: DEBUG nova.virt.hardware [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 819.922034] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080cd203-ce03-4836-a7f3-2c18d7078b62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.931996] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2522e9cf-2d6f-4b0b-b531-54e610d1c570 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.051720] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "f694fcb6-053b-4649-ac63-7fa98b1373eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 820.052016] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.052238] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "f694fcb6-053b-4649-ac63-7fa98b1373eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 820.052416] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.052575] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.055093] env[65121]: INFO nova.compute.manager [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Terminating instance [ 820.059083] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.210s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.061872] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.578s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.061872] env[65121]: DEBUG nova.objects.instance [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lazy-loading 'resources' on Instance uuid c1004635-b318-489c-9e16-6cb545279953 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.086036] env[65121]: INFO nova.scheduler.client.report [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Deleted allocations for instance e2d32a5e-c350-4b2b-9243-c3b412193a82 [ 820.139398] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "refresh_cache-75bf0b85-db28-4efa-a517-d7256873b09d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.139398] env[65121]: DEBUG nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Instance network_info: |[{"id": "6e8b4013-341c-4478-b8a7-1b7c43349ad8", "address": "fa:16:3e:98:fc:60", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e8b4013-34", "ovs_interfaceid": "6e8b4013-341c-4478-b8a7-1b7c43349ad8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 820.139951] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:fc:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e8b4013-341c-4478-b8a7-1b7c43349ad8', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.147747] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 820.149145] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.149395] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a82d247e-bad9-455b-ac4a-ffb5c6b455bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.178617] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.178617] env[65121]: value = "task-5106486" [ 820.178617] env[65121]: _type = "Task" [ 820.178617] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.188379] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106486, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.421948] env[65121]: INFO nova.compute.manager [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Took 45.14 seconds to build instance. [ 820.558055] env[65121]: DEBUG nova.network.neutron [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Successfully updated port: 30bb9dd1-838c-45e3-80f2-580a96b705ff {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 820.563372] env[65121]: DEBUG nova.compute.manager [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 820.563876] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.568054] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a191f2d-f99b-41a6-a76c-24e735d6e4fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.585752] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 820.588106] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b2a9f5c-d8d3-4ee3-a970-79ae81ead3e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.595446] env[65121]: DEBUG oslo_concurrency.lockutils [None req-16451e5a-37d2-4465-9998-f7d876f2a4ad tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "e2d32a5e-c350-4b2b-9243-c3b412193a82" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.944s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.601869] env[65121]: DEBUG oslo_vmware.api [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 820.601869] env[65121]: value = "task-5106487" [ 820.601869] env[65121]: _type = "Task" [ 820.601869] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.614317] env[65121]: DEBUG oslo_vmware.api [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.693485] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106486, 'name': CreateVM_Task, 'duration_secs': 0.407948} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.693583] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 820.694204] env[65121]: WARNING neutronclient.v2_0.client [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 820.694862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.694862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 820.695127] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 820.695536] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed90ab22-5c40-48b7-a927-2ffb87726265 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.704178] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 820.704178] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522b37fd-1b90-79ed-3619-3efa1d9b90d3" [ 820.704178] env[65121]: _type = "Task" [ 820.704178] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.714825] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522b37fd-1b90-79ed-3619-3efa1d9b90d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.926312] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bb10622-c3d0-44ef-b8aa-bd149a0b86d9 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.430s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.062950] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "refresh_cache-3f3bec5b-2834-497c-a454-a152cb992309" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.062950] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "refresh_cache-3f3bec5b-2834-497c-a454-a152cb992309" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 821.063206] env[65121]: DEBUG nova.network.neutron [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 821.118467] env[65121]: DEBUG oslo_vmware.api [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106487, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.143075] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2b37e0-44a9-47ee-8f76-e4d93d6bc212 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.151835] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b62b0f-379a-47a7-88fa-b4f9658301c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.190366] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b973c9c8-8fba-4fb2-9329-d455ff4e9a71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.202388] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb69439-5eaf-4cda-8811-0917fa91285e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.222922] env[65121]: DEBUG nova.compute.provider_tree [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.229112] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522b37fd-1b90-79ed-3619-3efa1d9b90d3, 'name': SearchDatastore_Task, 'duration_secs': 0.015068} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.229330] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 821.230035] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.230035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.230035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 821.230220] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.230384] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88988841-326e-403d-8b4a-0c0453c42ab4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.240836] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.241049] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 821.241810] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65350568-b489-4310-b58d-a73124369a48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.248506] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 821.248506] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ebac49-cd1a-7e0e-5656-7dc09b7696b4" [ 821.248506] env[65121]: _type = "Task" [ 821.248506] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.257938] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ebac49-cd1a-7e0e-5656-7dc09b7696b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.429479] env[65121]: DEBUG nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 821.566642] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.567045] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.616074] env[65121]: DEBUG oslo_vmware.api [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106487, 'name': PowerOffVM_Task, 'duration_secs': 0.529771} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.616239] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 821.616402] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 821.616653] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1c2d13a-fe4c-47b6-ae05-60338840f35d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.622235] env[65121]: DEBUG nova.network.neutron [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 821.670192] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.670192] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.699380] env[65121]: DEBUG nova.compute.manager [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 821.699910] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888376f9-2634-48c9-9de6-18e87d35be87 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.702973] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 821.703248] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 821.703535] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Deleting the datastore file [datastore2] f694fcb6-053b-4649-ac63-7fa98b1373eb {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 821.708933] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8005ab6a-d7b1-4f98-804e-2331e0371df7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.720841] env[65121]: DEBUG oslo_vmware.api [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for the task: (returnval){ [ 821.720841] env[65121]: value = "task-5106489" [ 821.720841] env[65121]: _type = "Task" [ 821.720841] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.732159] env[65121]: DEBUG nova.scheduler.client.report [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 821.735379] env[65121]: DEBUG oslo_vmware.api [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.770458] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ebac49-cd1a-7e0e-5656-7dc09b7696b4, 'name': SearchDatastore_Task, 'duration_secs': 0.021702} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.771854] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5f679fd-7461-42a1-954c-5a6a4cf9fa09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.776201] env[65121]: WARNING neutronclient.v2_0.client [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 821.776899] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.777427] env[65121]: WARNING openstack [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.792709] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 821.792709] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fb0d64-9beb-ca2b-f8bb-13dee0c9d2e5" [ 821.792709] env[65121]: _type = "Task" [ 821.792709] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.805871] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fb0d64-9beb-ca2b-f8bb-13dee0c9d2e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.849656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.849962] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.959649] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.989346] env[65121]: DEBUG nova.network.neutron [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Updating instance_info_cache with network_info: [{"id": "30bb9dd1-838c-45e3-80f2-580a96b705ff", "address": "fa:16:3e:9b:d6:e1", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb9dd1-83", "ovs_interfaceid": "30bb9dd1-838c-45e3-80f2-580a96b705ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 822.226503] env[65121]: INFO nova.compute.manager [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] instance snapshotting [ 822.227128] env[65121]: DEBUG nova.objects.instance [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'flavor' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.234752] env[65121]: DEBUG oslo_vmware.api [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Task: {'id': task-5106489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315299} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.235053] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.235375] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 822.235484] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.235546] env[65121]: INFO nova.compute.manager [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Took 1.67 seconds to destroy the instance on the hypervisor. [ 822.235778] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 822.235975] env[65121]: DEBUG nova.compute.manager [-] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 822.236176] env[65121]: DEBUG nova.network.neutron [-] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 822.236799] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 822.237015] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.237495] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 822.245364] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.184s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.247399] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.390s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.247969] env[65121]: DEBUG nova.objects.instance [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'resources' on Instance uuid 75114f97-fe50-4624-9333-303e411529ea {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.270840] env[65121]: INFO nova.scheduler.client.report [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleted allocations for instance c1004635-b318-489c-9e16-6cb545279953 [ 822.304646] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fb0d64-9beb-ca2b-f8bb-13dee0c9d2e5, 'name': SearchDatastore_Task, 'duration_secs': 0.015066} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.304916] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.305968] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 75bf0b85-db28-4efa-a517-d7256873b09d/75bf0b85-db28-4efa-a517-d7256873b09d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 822.305968] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8feac60-9bf0-4016-b5d0-5adec6dab09b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.314926] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 822.314926] env[65121]: value = "task-5106490" [ 822.314926] env[65121]: _type = "Task" [ 822.314926] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.316343] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 822.331906] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.372370] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Received event network-vif-plugged-6e8b4013-341c-4478-b8a7-1b7c43349ad8 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 822.372675] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Acquiring lock "75bf0b85-db28-4efa-a517-d7256873b09d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.372993] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Lock "75bf0b85-db28-4efa-a517-d7256873b09d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.373214] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Lock "75bf0b85-db28-4efa-a517-d7256873b09d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.373385] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] No waiting events found dispatching network-vif-plugged-6e8b4013-341c-4478-b8a7-1b7c43349ad8 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 822.373542] env[65121]: WARNING nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Received unexpected event network-vif-plugged-6e8b4013-341c-4478-b8a7-1b7c43349ad8 for instance with vm_state building and task_state spawning. [ 822.373793] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Received event network-changed-6e8b4013-341c-4478-b8a7-1b7c43349ad8 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 822.373835] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Refreshing instance network info cache due to event network-changed-6e8b4013-341c-4478-b8a7-1b7c43349ad8. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 822.374095] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Acquiring lock "refresh_cache-75bf0b85-db28-4efa-a517-d7256873b09d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.374178] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Acquired lock "refresh_cache-75bf0b85-db28-4efa-a517-d7256873b09d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.374362] env[65121]: DEBUG nova.network.neutron [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Refreshing network info cache for port 6e8b4013-341c-4478-b8a7-1b7c43349ad8 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 822.493193] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "refresh_cache-3f3bec5b-2834-497c-a454-a152cb992309" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.493620] env[65121]: DEBUG nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Instance network_info: |[{"id": "30bb9dd1-838c-45e3-80f2-580a96b705ff", "address": "fa:16:3e:9b:d6:e1", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb9dd1-83", "ovs_interfaceid": "30bb9dd1-838c-45e3-80f2-580a96b705ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 822.494110] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:d6:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30bb9dd1-838c-45e3-80f2-580a96b705ff', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.503243] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 822.503542] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.503778] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d3eb940-edbe-4c7d-8080-b1297b7dcfdf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.527592] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.527592] env[65121]: value = "task-5106491" [ 822.527592] env[65121]: _type = "Task" [ 822.527592] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.543850] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106491, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.682343] env[65121]: DEBUG nova.compute.manager [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 822.684033] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a018363b-b401-46b4-94c5-57e04c66233e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.735095] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b89cde0-3210-402c-9efa-31876e799d7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.761908] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b131be-3ddf-4600-8315-655b35689365 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.782061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-65df3c0d-fc13-4e69-8eb7-9af91690e3c6 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c1004635-b318-489c-9e16-6cb545279953" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.828s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.832784] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495713} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.833165] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 75bf0b85-db28-4efa-a517-d7256873b09d/75bf0b85-db28-4efa-a517-d7256873b09d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.833480] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.833844] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f176c989-0399-4bdb-a5f9-64dcd07a9ffe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.846468] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 822.846468] env[65121]: value = "task-5106492" [ 822.846468] env[65121]: _type = "Task" [ 822.846468] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.858092] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106492, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.877626] env[65121]: WARNING neutronclient.v2_0.client [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 822.878538] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.878718] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.047309] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106491, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.052371] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.052752] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.064153] env[65121]: DEBUG nova.network.neutron [-] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 823.149824] env[65121]: WARNING neutronclient.v2_0.client [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.150884] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.151396] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.198608] env[65121]: INFO nova.compute.manager [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] instance snapshotting [ 823.201483] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762f36cd-20db-49b6-a485-a26c70435b03 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.234302] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3af98c-2d23-490d-803f-b7c69d1704c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.275690] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 823.276192] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8c2b4019-a8f8-459a-9412-7f5a8493d87e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.281918] env[65121]: DEBUG nova.network.neutron [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Updated VIF entry in instance network info cache for port 6e8b4013-341c-4478-b8a7-1b7c43349ad8. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 823.282277] env[65121]: DEBUG nova.network.neutron [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Updating instance_info_cache with network_info: [{"id": "6e8b4013-341c-4478-b8a7-1b7c43349ad8", "address": "fa:16:3e:98:fc:60", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e8b4013-34", "ovs_interfaceid": "6e8b4013-341c-4478-b8a7-1b7c43349ad8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 823.285032] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 823.285032] env[65121]: value = "task-5106493" [ 823.285032] env[65121]: _type = "Task" [ 823.285032] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.299188] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106493, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.357438] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106492, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088437} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.360438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.361988] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0804f3ee-1213-4a60-b924-b05848015065 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.386239] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 75bf0b85-db28-4efa-a517-d7256873b09d/75bf0b85-db28-4efa-a517-d7256873b09d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.389291] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-859f14e0-81eb-4cbf-9cb1-dc46aa09956f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.404545] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.404642] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.404861] env[65121]: INFO nova.compute.manager [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Rebooting instance [ 823.415042] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 823.415042] env[65121]: value = "task-5106494" [ 823.415042] env[65121]: _type = "Task" [ 823.415042] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.424111] env[65121]: DEBUG nova.compute.manager [req-3e044ea8-f490-4375-8e6f-684c2b135098 req-8506a0ca-879c-4fb3-9580-fccef8e05a0d service nova] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Received event network-vif-deleted-203dc15c-4972-4c3d-9e50-af138f9134fe {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 823.431403] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106494, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.468243] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b31686-7097-4fb6-9acd-aaae76466ac1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.477335] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f91b270-0fb2-4eb1-bb90-f77be4795cf7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.513696] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85ab3e2-674a-4b2b-b4ae-e638232e6eac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.521949] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f36c0d-9ebe-4956-8bd1-f05c962baa2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.536981] env[65121]: DEBUG nova.compute.provider_tree [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.553254] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106491, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.567228] env[65121]: INFO nova.compute.manager [-] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Took 1.33 seconds to deallocate network for instance. [ 823.756661] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 823.757369] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9bf195cd-869c-4f35-bac5-c5271b0c2e7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.767527] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 823.767527] env[65121]: value = "task-5106495" [ 823.767527] env[65121]: _type = "Task" [ 823.767527] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.777716] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106495, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.786679] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Releasing lock "refresh_cache-75bf0b85-db28-4efa-a517-d7256873b09d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 823.787030] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Received event network-vif-plugged-30bb9dd1-838c-45e3-80f2-580a96b705ff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 823.787269] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Acquiring lock "3f3bec5b-2834-497c-a454-a152cb992309-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 823.787491] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Lock "3f3bec5b-2834-497c-a454-a152cb992309-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.787645] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Lock "3f3bec5b-2834-497c-a454-a152cb992309-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.787805] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] No waiting events found dispatching network-vif-plugged-30bb9dd1-838c-45e3-80f2-580a96b705ff {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 823.787966] env[65121]: WARNING nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Received unexpected event network-vif-plugged-30bb9dd1-838c-45e3-80f2-580a96b705ff for instance with vm_state building and task_state spawning. [ 823.788138] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Received event network-changed-30bb9dd1-838c-45e3-80f2-580a96b705ff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 823.788365] env[65121]: DEBUG nova.compute.manager [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Refreshing instance network info cache due to event network-changed-30bb9dd1-838c-45e3-80f2-580a96b705ff. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 823.788458] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Acquiring lock "refresh_cache-3f3bec5b-2834-497c-a454-a152cb992309" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.788586] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Acquired lock "refresh_cache-3f3bec5b-2834-497c-a454-a152cb992309" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 823.788775] env[65121]: DEBUG nova.network.neutron [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Refreshing network info cache for port 30bb9dd1-838c-45e3-80f2-580a96b705ff {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 823.801849] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106493, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.927125] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106494, 'name': ReconfigVM_Task, 'duration_secs': 0.309385} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.928758] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.929131] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquired lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 823.929300] env[65121]: DEBUG nova.network.neutron [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 823.931021] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 75bf0b85-db28-4efa-a517-d7256873b09d/75bf0b85-db28-4efa-a517-d7256873b09d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.932644] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-093957d1-d24e-4958-8951-fe4b95d4aeb9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.945426] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 823.945426] env[65121]: value = "task-5106496" [ 823.945426] env[65121]: _type = "Task" [ 823.945426] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.967628] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106496, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.047435] env[65121]: DEBUG nova.scheduler.client.report [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 824.050842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.051228] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.051297] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.051420] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.051561] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 824.052991] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106491, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.054837] env[65121]: INFO nova.compute.manager [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Terminating instance [ 824.074296] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.082648] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "274d0ccd-c707-4a68-b280-16de2bc74d73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.084592] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.237784] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.237895] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.279553] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106495, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.295576] env[65121]: WARNING neutronclient.v2_0.client [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.296348] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.297275] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.316518] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106493, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.404760] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.405025] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.433649] env[65121]: WARNING neutronclient.v2_0.client [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.434644] env[65121]: WARNING openstack [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.434762] env[65121]: WARNING openstack [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.455916] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106496, 'name': Rename_Task, 'duration_secs': 0.180869} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.460022] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 824.460022] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc5bfc97-457c-486b-8fd2-59cab4896bd4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.466026] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 824.466026] env[65121]: value = "task-5106497" [ 824.466026] env[65121]: _type = "Task" [ 824.466026] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.474274] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.489147] env[65121]: WARNING neutronclient.v2_0.client [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.489978] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.490454] env[65121]: WARNING openstack [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.553076] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106491, 'name': CreateVM_Task, 'duration_secs': 1.592505} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.555984] env[65121]: WARNING openstack [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.556555] env[65121]: WARNING openstack [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.566097] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.319s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 824.568878] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.569725] env[65121]: DEBUG nova.compute.manager [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 824.570058] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 824.576160] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 38.323s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.578461] env[65121]: WARNING neutronclient.v2_0.client [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.578987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.579265] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.579661] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 824.581043] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb68bbc-12f0-46bf-959e-33a4b2c5a056 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.588019] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50ac6a4a-fcca-44d7-8a38-9ba493243f84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.596028] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 824.596028] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5203d37f-9860-df80-cc42-eae29480f14a" [ 824.596028] env[65121]: _type = "Task" [ 824.596028] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.598463] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 824.610224] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdc4be00-03bc-438d-9c4f-11d3116d4177 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.622606] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5203d37f-9860-df80-cc42-eae29480f14a, 'name': SearchDatastore_Task, 'duration_secs': 0.016124} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.624178] env[65121]: INFO nova.scheduler.client.report [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted allocations for instance 75114f97-fe50-4624-9333-303e411529ea [ 824.627251] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.627781] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.628195] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.628481] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.629327] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.633222] env[65121]: DEBUG oslo_vmware.api [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 824.633222] env[65121]: value = "task-5106498" [ 824.633222] env[65121]: _type = "Task" [ 824.633222] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.633628] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a213b67-9e73-4d35-a785-44aa26638c4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.648702] env[65121]: DEBUG oslo_vmware.api [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.651038] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.651038] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 824.651275] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1748b65e-ab38-42e1-93d8-052229f072bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.658947] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 824.658947] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5272e70a-a884-6a40-44b5-2531c36d8036" [ 824.658947] env[65121]: _type = "Task" [ 824.658947] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.667366] env[65121]: WARNING neutronclient.v2_0.client [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 824.668222] env[65121]: WARNING openstack [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 824.668757] env[65121]: WARNING openstack [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 824.680363] env[65121]: DEBUG nova.network.neutron [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Updated VIF entry in instance network info cache for port 30bb9dd1-838c-45e3-80f2-580a96b705ff. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 824.680756] env[65121]: DEBUG nova.network.neutron [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Updating instance_info_cache with network_info: [{"id": "30bb9dd1-838c-45e3-80f2-580a96b705ff", "address": "fa:16:3e:9b:d6:e1", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bb9dd1-83", "ovs_interfaceid": "30bb9dd1-838c-45e3-80f2-580a96b705ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 824.689751] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5272e70a-a884-6a40-44b5-2531c36d8036, 'name': SearchDatastore_Task, 'duration_secs': 0.02139} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.690724] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcd9cdb0-c962-4da2-8ab2-8d40b7bd46c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.698563] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 824.698563] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5286b389-9621-da56-15e8-77cf7c8d5421" [ 824.698563] env[65121]: _type = "Task" [ 824.698563] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.710462] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5286b389-9621-da56-15e8-77cf7c8d5421, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.779605] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106495, 'name': CreateSnapshot_Task, 'duration_secs': 0.925054} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.779985] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 824.780918] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfd9546-26b2-4bf1-a33f-5d06dea2017d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.786822] env[65121]: DEBUG nova.network.neutron [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Updating instance_info_cache with network_info: [{"id": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "address": "fa:16:3e:af:f0:fd", "network": {"id": "f16ea9d5-3016-4746-8747-20bc1a0d5a8e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-801656862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9209f0dd88bc43129ae91f27d8212239", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab950b26-0d", "ovs_interfaceid": "ab950b26-0d72-483b-87f5-4686e5bd3c88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 824.807050] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106493, 'name': CreateSnapshot_Task, 'duration_secs': 1.374935} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.807429] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 824.808227] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d8c9e1-e9f7-4f30-8cce-0d2da4b3322e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.975997] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106497, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.089141] env[65121]: INFO nova.compute.claims [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.145525] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0138b4d2-99b4-40f6-825d-04bdf1e75ef0 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "75114f97-fe50-4624-9333-303e411529ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.519s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.153722] env[65121]: DEBUG oslo_vmware.api [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106498, 'name': PowerOffVM_Task, 'duration_secs': 0.209311} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.154190] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 825.154268] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 825.154703] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d30e1d1c-a194-48dc-9e48-3d11641e4291 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.183984] env[65121]: DEBUG oslo_concurrency.lockutils [req-884ee3cf-d8ee-4401-9aa1-c11692443079 req-02fc7464-1c68-4ff4-953d-be1ba76664e5 service nova] Releasing lock "refresh_cache-3f3bec5b-2834-497c-a454-a152cb992309" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.210706] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5286b389-9621-da56-15e8-77cf7c8d5421, 'name': SearchDatastore_Task, 'duration_secs': 0.011118} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.211608] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.211608] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 3f3bec5b-2834-497c-a454-a152cb992309/3f3bec5b-2834-497c-a454-a152cb992309.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.211608] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4cade6c-0325-435f-8007-8789163d7c24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.220208] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 825.220208] env[65121]: value = "task-5106500" [ 825.220208] env[65121]: _type = "Task" [ 825.220208] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.228225] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 825.228420] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 825.228612] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleting the datastore file [datastore1] b1ad9d15-7ef4-404a-9751-ecdd02ea768b {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.231703] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5242dfa2-95b6-4909-aab0-004d5e73d2ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.234379] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.242159] env[65121]: DEBUG oslo_vmware.api [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 825.242159] env[65121]: value = "task-5106501" [ 825.242159] env[65121]: _type = "Task" [ 825.242159] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.252798] env[65121]: DEBUG oslo_vmware.api [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.290316] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Releasing lock "refresh_cache-ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.308591] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 825.309440] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-044abc8f-c7e7-4a0b-b404-080ea62b7d27 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.329021] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 825.329021] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 825.329021] env[65121]: value = "task-5106502" [ 825.329021] env[65121]: _type = "Task" [ 825.329021] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.329386] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-19b6a9e3-39ce-4244-ba94-41a0c0b7e929 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.342169] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106502, 'name': CloneVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.344053] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 825.344053] env[65121]: value = "task-5106503" [ 825.344053] env[65121]: _type = "Task" [ 825.344053] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.353836] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106503, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.478307] env[65121]: DEBUG oslo_vmware.api [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106497, 'name': PowerOnVM_Task, 'duration_secs': 0.709538} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.478517] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.478721] env[65121]: INFO nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Took 8.45 seconds to spawn the instance on the hypervisor. [ 825.478896] env[65121]: DEBUG nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 825.479891] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78314a5c-ec57-4134-adbb-6db99f3c12d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.601242] env[65121]: INFO nova.compute.resource_tracker [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating resource usage from migration 2ce31b97-0aef-4465-995a-13b22493087b [ 825.737776] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106500, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.755344] env[65121]: DEBUG oslo_vmware.api [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213387} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.755796] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.755977] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 825.756183] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.756354] env[65121]: INFO nova.compute.manager [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Took 1.19 seconds to destroy the instance on the hypervisor. [ 825.756604] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 825.756841] env[65121]: DEBUG nova.compute.manager [-] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 825.756954] env[65121]: DEBUG nova.network.neutron [-] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 825.757619] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.758195] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.758329] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.801169] env[65121]: DEBUG nova.compute.manager [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 825.801410] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d0b761-e059-460b-a694-68d3c00c91b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.844793] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106502, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.859201] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106503, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.872088] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 826.002366] env[65121]: INFO nova.compute.manager [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Took 46.38 seconds to build instance. [ 826.236191] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532981} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.236191] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 3f3bec5b-2834-497c-a454-a152cb992309/3f3bec5b-2834-497c-a454-a152cb992309.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 826.236358] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.236537] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-594d06a3-b970-4453-a6df-b98f551a6af8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.245997] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 826.245997] env[65121]: value = "task-5106504" [ 826.245997] env[65121]: _type = "Task" [ 826.245997] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.265091] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106504, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.344138] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106502, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.360913] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106503, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.422616] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fa2847-6e39-4e83-8d75-e53c332927f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.434542] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd79f7ac-c12a-454b-98fb-abbe914980e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.473114] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b513793-730e-47d0-b921-46a59f0f07b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.478722] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "df5abcb5-583f-4b28-a074-3a3221d74d87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.478722] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.485186] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d92149a-d060-4e5b-85aa-c7f659618510 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.500663] env[65121]: DEBUG nova.compute.provider_tree [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.505390] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18fafa7e-e5d1-433c-945d-3654d088feff tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75bf0b85-db28-4efa-a517-d7256873b09d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.889s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.671184] env[65121]: DEBUG nova.network.neutron [-] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 826.760828] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106504, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081913} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.762528] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.767758] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f847afd-fcc0-4a2d-96b1-8394df8cec42 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.798762] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 3f3bec5b-2834-497c-a454-a152cb992309/3f3bec5b-2834-497c-a454-a152cb992309.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.799156] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb20ce26-f425-4f1a-ac2a-f3a697d82d24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.822100] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d46725e-e165-41e8-a509-730084d0288c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.825643] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 826.825643] env[65121]: value = "task-5106505" [ 826.825643] env[65121]: _type = "Task" [ 826.825643] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.831522] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Doing hard reboot of VM {{(pid=65121) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 826.831993] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-e747d97c-32b3-47d5-8322-1d9bf3ef4643 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.838123] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106505, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.842596] env[65121]: DEBUG oslo_vmware.api [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 826.842596] env[65121]: value = "task-5106506" [ 826.842596] env[65121]: _type = "Task" [ 826.842596] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.850589] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106502, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.858283] env[65121]: DEBUG oslo_vmware.api [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106506, 'name': ResetVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.866804] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106503, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.940701] env[65121]: DEBUG nova.compute.manager [req-84d7ea8c-c07a-40db-8e71-fec586ef88a9 req-bc92814f-1052-4dcd-b86d-5b5f9771139f service nova] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Received event network-vif-deleted-95f5dff8-7aaa-4dd4-b724-b8f5704e7eff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 827.004699] env[65121]: DEBUG nova.scheduler.client.report [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 827.011267] env[65121]: DEBUG nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 827.174086] env[65121]: INFO nova.compute.manager [-] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Took 1.42 seconds to deallocate network for instance. [ 827.337830] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.359335] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106502, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.359861] env[65121]: DEBUG oslo_vmware.api [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106506, 'name': ResetVM_Task, 'duration_secs': 0.121846} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.364960] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Did hard reboot of VM {{(pid=65121) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 827.365530] env[65121]: DEBUG nova.compute.manager [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 827.366578] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ac57ee-1cce-4bba-ac90-1a0c54d19c16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.379394] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106503, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.516842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.941s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.517070] env[65121]: INFO nova.compute.manager [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Migrating [ 827.527224] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.627s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.529883] env[65121]: INFO nova.compute.claims [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.565410] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.681837] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.837590] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106505, 'name': ReconfigVM_Task, 'duration_secs': 0.681605} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.837834] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 3f3bec5b-2834-497c-a454-a152cb992309/3f3bec5b-2834-497c-a454-a152cb992309.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.838629] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-748615a8-bd75-4705-ace3-09d9f7d6d3c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.851882] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106502, 'name': CloneVM_Task, 'duration_secs': 2.310397} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.854125] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Created linked-clone VM from snapshot [ 827.854125] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 827.854125] env[65121]: value = "task-5106507" [ 827.854125] env[65121]: _type = "Task" [ 827.854125] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.855239] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e9025d-a125-4303-9769-8ad8c73ede92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.879964] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Uploading image 790825ff-9344-4100-b5fb-4721ce889ec7 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 827.882581] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106507, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.889951] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106503, 'name': CloneVM_Task, 'duration_secs': 2.355664} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.890445] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Created linked-clone VM from snapshot [ 827.891445] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76c7415-8500-4e9a-a3ba-ca5a1384bf0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.894976] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2490f07-1bc3-4c74-87f4-49af79480326 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.490s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.903076] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Uploading image 4420ebad-90bb-4083-ad61-fa47fb6880b2 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 827.907305] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 827.907700] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9c0c56f0-e4f9-4aca-8c02-96c092202594 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.919883] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 827.919883] env[65121]: value = "task-5106508" [ 827.919883] env[65121]: _type = "Task" [ 827.919883] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.933778] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106508, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.936535] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 827.936535] env[65121]: value = "vm-993429" [ 827.936535] env[65121]: _type = "VirtualMachine" [ 827.936535] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 827.936885] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0d9f35a4-867d-40f0-8e5b-da3338bcea81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.946059] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease: (returnval){ [ 827.946059] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f987d8-2bec-3367-315c-49506b03b659" [ 827.946059] env[65121]: _type = "HttpNfcLease" [ 827.946059] env[65121]: } obtained for exporting VM: (result){ [ 827.946059] env[65121]: value = "vm-993429" [ 827.946059] env[65121]: _type = "VirtualMachine" [ 827.946059] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 827.946493] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the lease: (returnval){ [ 827.946493] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f987d8-2bec-3367-315c-49506b03b659" [ 827.946493] env[65121]: _type = "HttpNfcLease" [ 827.946493] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 827.958967] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 827.958967] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f987d8-2bec-3367-315c-49506b03b659" [ 827.958967] env[65121]: _type = "HttpNfcLease" [ 827.958967] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 828.056563] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.056563] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.056563] env[65121]: DEBUG nova.network.neutron [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 828.373629] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106507, 'name': Rename_Task, 'duration_secs': 0.177951} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.373629] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.373858] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd25749a-6c10-44c9-aadd-1def0cedd986 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.383267] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 828.383267] env[65121]: value = "task-5106510" [ 828.383267] env[65121]: _type = "Task" [ 828.383267] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.395880] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.436650] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106508, 'name': Destroy_Task, 'duration_secs': 0.38487} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.436748] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Destroyed the VM [ 828.436938] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 828.440566] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2e0a2a5f-e57e-4f92-a014-cdebcdb8873d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.447214] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 828.447214] env[65121]: value = "task-5106511" [ 828.447214] env[65121]: _type = "Task" [ 828.447214] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.463168] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106511, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.463443] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 828.463443] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f987d8-2bec-3367-315c-49506b03b659" [ 828.463443] env[65121]: _type = "HttpNfcLease" [ 828.463443] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 828.463757] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 828.463757] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f987d8-2bec-3367-315c-49506b03b659" [ 828.463757] env[65121]: _type = "HttpNfcLease" [ 828.463757] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 828.464727] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50dbee0b-1bd4-46f5-be05-18622acb53cc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.474397] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb7fc5-3096-8d91-ded2-2af022b52965/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 828.474771] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb7fc5-3096-8d91-ded2-2af022b52965/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 828.552382] env[65121]: WARNING neutronclient.v2_0.client [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.553244] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.553701] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.581548] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1cf72c48-58ca-4715-a171-11d45ce1beab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.770412] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.770412] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.879452] env[65121]: WARNING neutronclient.v2_0.client [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.880187] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.880756] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.904323] env[65121]: DEBUG oslo_vmware.api [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106510, 'name': PowerOnVM_Task, 'duration_secs': 0.521774} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.904445] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.904666] env[65121]: INFO nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Took 9.03 seconds to spawn the instance on the hypervisor. [ 828.904837] env[65121]: DEBUG nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 828.908789] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9e2dfa-912d-4aae-9921-346de1c6b699 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.921078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "75bf0b85-db28-4efa-a517-d7256873b09d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.921078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75bf0b85-db28-4efa-a517-d7256873b09d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.921078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "75bf0b85-db28-4efa-a517-d7256873b09d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.921078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75bf0b85-db28-4efa-a517-d7256873b09d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.921078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75bf0b85-db28-4efa-a517-d7256873b09d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.925604] env[65121]: INFO nova.compute.manager [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Terminating instance [ 828.964249] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106511, 'name': RemoveSnapshot_Task} progress is 76%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.125280] env[65121]: DEBUG nova.network.neutron [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance_info_cache with network_info: [{"id": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "address": "fa:16:3e:c8:d8:39", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79bfb570-6e", "ovs_interfaceid": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 829.359568] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c14a5bd-0c4a-48d1-b681-170fab9347a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.368834] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84cc4d2-df47-4297-aceb-f0314e5763bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.407014] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ac90ad-1d9d-40f1-9396-adbc4083efbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.416828] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45650c1-40b3-487f-ac5a-bb580cca064b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.447472] env[65121]: DEBUG nova.compute.manager [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 829.447726] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.448930] env[65121]: DEBUG nova.compute.provider_tree [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.451422] env[65121]: INFO nova.compute.manager [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Took 49.79 seconds to build instance. [ 829.453995] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07847aef-e558-43fd-a35a-5730c4ec99c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.468249] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 829.474058] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7c5d061-f7ef-4dab-adc7-2286624d26da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.478081] env[65121]: DEBUG oslo_vmware.api [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106511, 'name': RemoveSnapshot_Task, 'duration_secs': 0.714628} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.478081] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 829.489889] env[65121]: DEBUG oslo_vmware.api [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 829.489889] env[65121]: value = "task-5106512" [ 829.489889] env[65121]: _type = "Task" [ 829.489889] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.502879] env[65121]: DEBUG oslo_vmware.api [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106512, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.630406] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.953968] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb33dd7b-e339-4eb2-8eaf-39a544cb2241 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "3f3bec5b-2834-497c-a454-a152cb992309" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.302s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.955210] env[65121]: DEBUG nova.scheduler.client.report [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 829.983308] env[65121]: WARNING nova.compute.manager [None req-a258204a-9631-4cb7-8808-9fd09c776b85 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Image not found during snapshot: nova.exception.ImageNotFound: Image 790825ff-9344-4100-b5fb-4721ce889ec7 could not be found. [ 830.007836] env[65121]: DEBUG oslo_vmware.api [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106512, 'name': PowerOffVM_Task, 'duration_secs': 0.345896} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.008265] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.008701] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 830.009310] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a94cb29-5378-41c8-81da-9daefc5a82b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.037469] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.038066] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.041396] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 830.041689] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.003s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.042035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.047104] env[65121]: INFO nova.compute.manager [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Terminating instance [ 830.083199] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.083470] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.083646] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleting the datastore file [datastore1] 75bf0b85-db28-4efa-a517-d7256873b09d {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.083924] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30d8ba7e-39af-4a1b-905c-342582be5e50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.094470] env[65121]: DEBUG oslo_vmware.api [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 830.094470] env[65121]: value = "task-5106514" [ 830.094470] env[65121]: _type = "Task" [ 830.094470] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.105220] env[65121]: DEBUG oslo_vmware.api [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106514, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.461531] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.934s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.462144] env[65121]: DEBUG nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 830.465192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.972s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.466774] env[65121]: INFO nova.compute.claims [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.472304] env[65121]: DEBUG nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 830.549825] env[65121]: DEBUG nova.compute.manager [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 830.550052] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 830.551095] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bc1624-d80e-40c2-8da4-cb6a4f99fc05 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.560337] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.561090] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2df61e85-124e-4902-9a59-9bb495f5d546 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.570611] env[65121]: DEBUG oslo_vmware.api [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 830.570611] env[65121]: value = "task-5106515" [ 830.570611] env[65121]: _type = "Task" [ 830.570611] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.582741] env[65121]: DEBUG oslo_vmware.api [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.605766] env[65121]: DEBUG oslo_vmware.api [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106514, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.241798} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.605836] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.606042] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 830.606781] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 830.606781] env[65121]: INFO nova.compute.manager [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 830.606781] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 830.607046] env[65121]: DEBUG nova.compute.manager [-] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 830.607144] env[65121]: DEBUG nova.network.neutron [-] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 830.607411] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.607940] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.608264] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.802278] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.978332] env[65121]: DEBUG nova.compute.utils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 830.983539] env[65121]: DEBUG nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 831.007622] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.084534] env[65121]: DEBUG oslo_vmware.api [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106515, 'name': PowerOffVM_Task, 'duration_secs': 0.317997} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.085280] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 831.085883] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 831.086201] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f817c22-93fe-4ca2-83b3-c025dbb1ae96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.155433] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b894b1d3-2102-48e8-9bcb-eb6e6382b45f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.183540] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance '293f93f2-c01d-42c8-b1a7-3056805c77de' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 831.189064] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 831.189835] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 831.189835] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Deleting the datastore file [datastore2] ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.190184] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01d740f9-dfbb-4de0-879d-c6060c44d481 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.198459] env[65121]: DEBUG oslo_vmware.api [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for the task: (returnval){ [ 831.198459] env[65121]: value = "task-5106517" [ 831.198459] env[65121]: _type = "Task" [ 831.198459] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.209913] env[65121]: DEBUG oslo_vmware.api [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106517, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.487168] env[65121]: DEBUG nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 831.556341] env[65121]: DEBUG nova.network.neutron [-] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 831.692271] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.694521] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc7cad56-db5d-4846-b0b2-c9cb7fd3ab3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.705723] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 831.705723] env[65121]: value = "task-5106518" [ 831.705723] env[65121]: _type = "Task" [ 831.705723] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.716435] env[65121]: DEBUG oslo_vmware.api [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Task: {'id': task-5106517, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.270419} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.717324] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.717639] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.717875] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.718099] env[65121]: INFO nova.compute.manager [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Took 1.17 seconds to destroy the instance on the hypervisor. [ 831.718415] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 831.718809] env[65121]: DEBUG nova.compute.manager [-] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 831.718809] env[65121]: DEBUG nova.network.neutron [-] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 831.719252] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.720071] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.720488] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.732775] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106518, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.784452] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.960849] env[65121]: DEBUG nova.compute.manager [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 831.961776] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e3e1af-18ea-4681-8d33-1d175e7d46af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.064165] env[65121]: INFO nova.compute.manager [-] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Took 1.46 seconds to deallocate network for instance. [ 832.076793] env[65121]: DEBUG nova.compute.manager [req-a39b5175-ebb0-446c-9fbf-d4b490044199 req-02595701-f644-49b1-8217-e80ff46f20bc service nova] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Received event network-vif-deleted-6e8b4013-341c-4478-b8a7-1b7c43349ad8 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 832.109577] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "90cfd88c-286e-4c87-bf27-909873359997" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.110044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "90cfd88c-286e-4c87-bf27-909873359997" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.110044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "90cfd88c-286e-4c87-bf27-909873359997-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.110228] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "90cfd88c-286e-4c87-bf27-909873359997-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.110421] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "90cfd88c-286e-4c87-bf27-909873359997-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 832.113069] env[65121]: INFO nova.compute.manager [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Terminating instance [ 832.228333] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106518, 'name': PowerOffVM_Task, 'duration_secs': 0.215128} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.228698] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.229214] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance '293f93f2-c01d-42c8-b1a7-3056805c77de' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 832.277763] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b147fc-2ff7-4d55-ae0c-c95f489c177c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.289893] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67840ff5-0e0e-4d67-a1d2-eaf11c3ba8fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.332599] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9a6ea8-b0a5-4192-a0c3-3d98df577620 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.342794] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5474541-468d-47bb-b0f3-2820aad36055 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.359575] env[65121]: DEBUG nova.compute.provider_tree [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.475170] env[65121]: DEBUG nova.compute.manager [req-002b89ce-c8cd-47cc-bc8a-d6c08fcd2ae5 req-f94447e6-177f-4dee-9614-2242323d8486 service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Received event network-vif-deleted-ab950b26-0d72-483b-87f5-4686e5bd3c88 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 832.475783] env[65121]: INFO nova.compute.manager [req-002b89ce-c8cd-47cc-bc8a-d6c08fcd2ae5 req-f94447e6-177f-4dee-9614-2242323d8486 service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Neutron deleted interface ab950b26-0d72-483b-87f5-4686e5bd3c88; detaching it from the instance and deleting it from the info cache [ 832.475783] env[65121]: DEBUG nova.network.neutron [req-002b89ce-c8cd-47cc-bc8a-d6c08fcd2ae5 req-f94447e6-177f-4dee-9614-2242323d8486 service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.477891] env[65121]: INFO nova.compute.manager [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] instance snapshotting [ 832.480275] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14402071-6e22-4cf6-91ba-565fe88b864c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.502847] env[65121]: DEBUG nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 832.506260] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fafb8ae-9752-4d4f-bbb1-9aa458325d3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.532532] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 832.532688] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 832.532802] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 832.532992] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 832.533192] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 832.533438] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 832.533656] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.533808] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 832.534049] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 832.535460] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 832.535460] env[65121]: DEBUG nova.virt.hardware [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 832.535460] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08649fdd-04b0-4eaa-ba24-b9a8f48461ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.547948] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b42ab96-9c3a-40de-be3b-57372ea25933 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.563813] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.569644] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Creating folder: Project (63aebdc51e41434a85385519496246e9). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.570778] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c59871c-3b95-4d4f-9cfc-ec48f6be6d1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.579232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.583828] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Created folder: Project (63aebdc51e41434a85385519496246e9) in parent group-v993268. [ 832.584111] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Creating folder: Instances. Parent ref: group-v993430. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.584411] env[65121]: DEBUG nova.network.neutron [-] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.585701] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e299406-8223-426d-8e85-4ba5d96ed45c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.597727] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Created folder: Instances in parent group-v993430. [ 832.598437] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 832.598638] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 832.598844] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73aece0b-2817-4a68-92a2-5208fa436f2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.617899] env[65121]: DEBUG nova.compute.manager [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 832.618110] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.619014] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d80e2e-a1db-4c31-a0be-54ed2511aa75 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.623156] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.623156] env[65121]: value = "task-5106521" [ 832.623156] env[65121]: _type = "Task" [ 832.623156] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.631135] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.631945] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e54ebd9-3a1e-4bf6-9911-b9fee5ca7f77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.637633] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106521, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.644383] env[65121]: DEBUG oslo_vmware.api [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 832.644383] env[65121]: value = "task-5106522" [ 832.644383] env[65121]: _type = "Task" [ 832.644383] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.655333] env[65121]: DEBUG oslo_vmware.api [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 832.739852] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 832.746273] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95a66c8a-f8fe-41cd-9a72-ddb7a13b93f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.764810] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 832.764810] env[65121]: value = "task-5106523" [ 832.764810] env[65121]: _type = "Task" [ 832.764810] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.778763] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106523, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.866336] env[65121]: DEBUG nova.scheduler.client.report [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 832.980648] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0d36440-e455-45d7-92a6-c4b81f81ba3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.991927] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03891331-8e43-4f28-ab92-60e990e490b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.038667] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 833.039313] env[65121]: DEBUG nova.compute.manager [req-002b89ce-c8cd-47cc-bc8a-d6c08fcd2ae5 req-f94447e6-177f-4dee-9614-2242323d8486 service nova] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Detach interface failed, port_id=ab950b26-0d72-483b-87f5-4686e5bd3c88, reason: Instance ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 833.039961] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5efe93ae-238c-4cd9-8ba0-5bdcedbb5788 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.049038] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 833.049038] env[65121]: value = "task-5106524" [ 833.049038] env[65121]: _type = "Task" [ 833.049038] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.058446] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106524, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.089499] env[65121]: INFO nova.compute.manager [-] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Took 1.37 seconds to deallocate network for instance. [ 833.135908] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106521, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.157051] env[65121]: DEBUG oslo_vmware.api [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106522, 'name': PowerOffVM_Task, 'duration_secs': 0.382556} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.157051] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.157051] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.157051] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5c5e004-5f70-4ae7-8a19-2237689d38b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.249152] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.250802] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.250802] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleting the datastore file [datastore2] 90cfd88c-286e-4c87-bf27-909873359997 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.250802] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd600d5c-4d31-4e32-9f27-15df3f33b8c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.259009] env[65121]: DEBUG oslo_vmware.api [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 833.259009] env[65121]: value = "task-5106526" [ 833.259009] env[65121]: _type = "Task" [ 833.259009] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.268580] env[65121]: DEBUG oslo_vmware.api [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.277881] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106523, 'name': ReconfigVM_Task, 'duration_secs': 0.250354} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.278540] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance '293f93f2-c01d-42c8-b1a7-3056805c77de' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 833.371875] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.907s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.372331] env[65121]: DEBUG nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 833.375331] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 44.396s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.375569] env[65121]: DEBUG nova.objects.instance [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lazy-loading 'resources' on Instance uuid aab9317b-4ee6-48b3-905b-859a5996f33d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.564354] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106524, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.598718] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.636637] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106521, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.771031] env[65121]: DEBUG oslo_vmware.api [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.478471} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.771031] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.771031] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.771402] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.771402] env[65121]: INFO nova.compute.manager [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Took 1.15 seconds to destroy the instance on the hypervisor. [ 833.771518] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 833.771745] env[65121]: DEBUG nova.compute.manager [-] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 833.771896] env[65121]: DEBUG nova.network.neutron [-] [instance: 90cfd88c-286e-4c87-bf27-909873359997] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 833.772246] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.772966] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.777021] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.784102] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:22:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='8a66939b-83da-49a6-8c3f-9ca9e786ade0',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1368792546',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 833.784381] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 833.784531] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 833.784759] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 833.784925] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 833.785267] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 833.785517] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.785721] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 833.786189] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 833.786189] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 833.786307] env[65121]: DEBUG nova.virt.hardware [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 833.792296] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfiguring VM instance instance-0000002a to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 833.792296] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6aad2544-1372-4757-b155-730dcab381a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.815402] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.820257] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 833.820257] env[65121]: value = "task-5106527" [ 833.820257] env[65121]: _type = "Task" [ 833.820257] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.833444] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106527, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.880701] env[65121]: DEBUG nova.compute.utils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 833.880701] env[65121]: DEBUG nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 833.881213] env[65121]: DEBUG nova.network.neutron [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 833.881716] env[65121]: WARNING neutronclient.v2_0.client [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.883152] env[65121]: WARNING neutronclient.v2_0.client [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.885021] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.885021] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.959174] env[65121]: DEBUG nova.policy [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10dd394ea47044e0ad78daacb6aa161a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad1aea30d62c45e193c5a54c429ce7ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 834.068045] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106524, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.138838] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106521, 'name': CreateVM_Task, 'duration_secs': 1.490844} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.139212] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.142698] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.142897] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.143338] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 834.143923] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77461f37-b95c-4835-a20c-779230f6b0a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.151326] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 834.151326] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528194a1-28a2-c912-6e1f-a8ef786fc787" [ 834.151326] env[65121]: _type = "Task" [ 834.151326] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.163163] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528194a1-28a2-c912-6e1f-a8ef786fc787, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.333714] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106527, 'name': ReconfigVM_Task, 'duration_secs': 0.459042} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.337031] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfigured VM instance instance-0000002a to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 834.337031] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103ba66b-4f8b-4996-bf76-d9f1ccfdaa26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.367252] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.371021] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b92f2bd5-aeb6-4535-b2a7-fed634f5c7e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.393622] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 834.393622] env[65121]: value = "task-5106528" [ 834.393622] env[65121]: _type = "Task" [ 834.393622] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.401686] env[65121]: DEBUG nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 834.415946] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.418175] env[65121]: DEBUG nova.network.neutron [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Successfully created port: 71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 834.563750] env[65121]: DEBUG nova.network.neutron [-] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 834.567029] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106524, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.672087] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528194a1-28a2-c912-6e1f-a8ef786fc787, 'name': SearchDatastore_Task, 'duration_secs': 0.013798} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.673323] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.673323] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.673323] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.673323] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.673323] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.673805] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e13682c5-3bde-4968-ab64-436c8d1b7671 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.686855] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.687057] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.687970] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4ae03d8-b0d5-4db8-949a-2e6d6ee04752 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.697245] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 834.697245] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523c8c78-3aa8-4021-3cb0-b3e2ced86876" [ 834.697245] env[65121]: _type = "Task" [ 834.697245] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.709994] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523c8c78-3aa8-4021-3cb0-b3e2ced86876, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.749852] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd675fa-f8f1-407d-ba26-ad16151f454a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.759859] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d17d04-8aaa-4b98-9eea-332bb7faa733 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.800761] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22db9c0-68b0-4b38-bd40-d49c4a0dfcbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.810727] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c251d24-64df-4a2b-87f6-017ba56e0637 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.817634] env[65121]: DEBUG nova.compute.manager [req-f25aab6f-1a8a-4c59-93f7-beb6e3076042 req-52d2ab60-3612-43da-9174-afe04bb1106c service nova] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Received event network-vif-deleted-9127a7ec-601e-4377-adfb-134eaaa9c000 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 834.829438] env[65121]: DEBUG nova.compute.provider_tree [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.905839] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106528, 'name': ReconfigVM_Task, 'duration_secs': 0.296608} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.906320] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.906519] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance '293f93f2-c01d-42c8-b1a7-3056805c77de' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 835.066187] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106524, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.069971] env[65121]: INFO nova.compute.manager [-] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Took 1.30 seconds to deallocate network for instance. [ 835.209024] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523c8c78-3aa8-4021-3cb0-b3e2ced86876, 'name': SearchDatastore_Task, 'duration_secs': 0.027073} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.210127] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1824500-31c2-4c29-aaa6-b94a1f3f4861 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.216618] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 835.216618] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52100f52-0e51-0b93-3b67-99a04dddef2c" [ 835.216618] env[65121]: _type = "Task" [ 835.216618] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.225310] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52100f52-0e51-0b93-3b67-99a04dddef2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.355254] env[65121]: ERROR nova.scheduler.client.report [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] [req-312c3097-2b4d-4f80-ba68-6002e0acd88c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-312c3097-2b4d-4f80-ba68-6002e0acd88c"}]} [ 835.372442] env[65121]: DEBUG nova.scheduler.client.report [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 835.388975] env[65121]: DEBUG nova.scheduler.client.report [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 835.389360] env[65121]: DEBUG nova.compute.provider_tree [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.404941] env[65121]: DEBUG nova.scheduler.client.report [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 835.413909] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02e4248-cd4b-4dce-b0eb-50ad4fb68c4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.436239] env[65121]: DEBUG nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 835.439274] env[65121]: DEBUG nova.scheduler.client.report [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 835.442202] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40347fb8-abcd-4ce0-a2da-04eec043be3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.464992] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance '293f93f2-c01d-42c8-b1a7-3056805c77de' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 835.481306] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 835.481575] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 835.481722] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 835.481892] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 835.482068] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 835.482228] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 835.482432] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.482580] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 835.482739] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 835.482892] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 835.483071] env[65121]: DEBUG nova.virt.hardware [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 835.483928] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c348052-374a-4500-8e29-e1598c6dc959 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.495023] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2bbde7-59c0-4a47-bbd4-d42b8bf870ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.566539] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106524, 'name': CreateSnapshot_Task, 'duration_secs': 2.092268} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.566932] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 835.567736] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858cd4c2-20b4-432b-9576-0b684ae2001e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.575724] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.733170] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52100f52-0e51-0b93-3b67-99a04dddef2c, 'name': SearchDatastore_Task, 'duration_secs': 0.025677} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.733170] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 835.733170] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.735923] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d256788-eb76-49c5-a211-39b857b0ed4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.744451] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 835.744451] env[65121]: value = "task-5106529" [ 835.744451] env[65121]: _type = "Task" [ 835.744451] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.754132] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.980981] env[65121]: WARNING neutronclient.v2_0.client [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.020344] env[65121]: DEBUG nova.network.neutron [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Port 79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 836.053938] env[65121]: DEBUG nova.network.neutron [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Successfully updated port: 71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 836.065194] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35e4080-de9c-40e2-bd63-ff29ad7341b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.075037] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9964891-c12a-4970-a215-05f349e5ea95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.121180] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 836.123405] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-000ba66f-64c7-4037-81fe-2dafba15a663 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.126101] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68dec95-be5e-42b0-a698-ec526fad31d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.135728] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cc1b56-0d98-456c-8648-827b60f86664 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.139852] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 836.139852] env[65121]: value = "task-5106530" [ 836.139852] env[65121]: _type = "Task" [ 836.139852] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.153271] env[65121]: DEBUG nova.compute.provider_tree [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.163947] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106530, 'name': CloneVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.259969] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106529, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.557217] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.557334] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 836.557589] env[65121]: DEBUG nova.network.neutron [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 836.652136] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106530, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.693108] env[65121]: DEBUG nova.scheduler.client.report [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 836.693108] env[65121]: DEBUG nova.compute.provider_tree [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 86 to 87 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 836.693108] env[65121]: DEBUG nova.compute.provider_tree [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.758417] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106529, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.833524} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.758417] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.758622] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.758867] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70191916-362d-46c2-bc41-9fbf073cf8df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.769090] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 836.769090] env[65121]: value = "task-5106531" [ 836.769090] env[65121]: _type = "Task" [ 836.769090] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.781097] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106531, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.874708] env[65121]: DEBUG nova.compute.manager [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-vif-plugged-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 836.874955] env[65121]: DEBUG oslo_concurrency.lockutils [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.876082] env[65121]: DEBUG oslo_concurrency.lockutils [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 836.876082] env[65121]: DEBUG oslo_concurrency.lockutils [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.876082] env[65121]: DEBUG nova.compute.manager [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] No waiting events found dispatching network-vif-plugged-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 836.876082] env[65121]: WARNING nova.compute.manager [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received unexpected event network-vif-plugged-71e0942f-5026-4128-ba81-16311feb9b3e for instance with vm_state building and task_state spawning. [ 836.876082] env[65121]: DEBUG nova.compute.manager [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 836.876082] env[65121]: DEBUG nova.compute.manager [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing instance network info cache due to event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 836.876631] env[65121]: DEBUG oslo_concurrency.lockutils [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Acquiring lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.044207] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.044768] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.045128] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.060560] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.061308] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.103964] env[65121]: DEBUG nova.network.neutron [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 837.135487] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.136458] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.153677] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106530, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.199957] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.825s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.203033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.696s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.204832] env[65121]: INFO nova.compute.claims [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.228262] env[65121]: WARNING neutronclient.v2_0.client [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 837.229098] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.229485] env[65121]: WARNING openstack [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.241139] env[65121]: INFO nova.scheduler.client.report [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Deleted allocations for instance aab9317b-4ee6-48b3-905b-859a5996f33d [ 837.283385] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106531, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076398} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.283833] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.284583] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b93d48-ff96-4d61-b186-8ad0b90ba836 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.308581] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.309309] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8cbc81e-5a9d-4506-a97b-9616d1ddd28b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.336531] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 837.336531] env[65121]: value = "task-5106532" [ 837.336531] env[65121]: _type = "Task" [ 837.336531] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.344256] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106532, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.389432] env[65121]: DEBUG nova.network.neutron [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e0942f-50", "ovs_interfaceid": "71e0942f-5026-4128-ba81-16311feb9b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.655813] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106530, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.752727] env[65121]: DEBUG oslo_concurrency.lockutils [None req-29389141-9df3-4f8b-a761-4943b62259a7 tempest-ServerAddressesNegativeTestJSON-2099705515 tempest-ServerAddressesNegativeTestJSON-2099705515-project-member] Lock "aab9317b-4ee6-48b3-905b-859a5996f33d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 52.574s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.847097] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106532, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.893924] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.893924] env[65121]: DEBUG nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Instance network_info: |[{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e0942f-50", "ovs_interfaceid": "71e0942f-5026-4128-ba81-16311feb9b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 837.893924] env[65121]: DEBUG oslo_concurrency.lockutils [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Acquired lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 837.893924] env[65121]: DEBUG nova.network.neutron [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 837.895690] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:a9:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a64108f9-df0a-4feb-bbb5-97f5841c356c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71e0942f-5026-4128-ba81-16311feb9b3e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.904297] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating folder: Project (ad1aea30d62c45e193c5a54c429ce7ba). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.905326] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd3bd189-9896-43bf-85c9-4ab3c51a0558 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.920644] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created folder: Project (ad1aea30d62c45e193c5a54c429ce7ba) in parent group-v993268. [ 837.922305] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating folder: Instances. Parent ref: group-v993435. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.922305] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6a672e7-271d-444c-bf8f-395e0a1baaee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.936952] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created folder: Instances in parent group-v993435. [ 837.937226] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 837.937435] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.937649] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a3bc4f0-d234-4df9-9eed-ebdebf6eece7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.961782] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.961782] env[65121]: value = "task-5106535" [ 837.961782] env[65121]: _type = "Task" [ 837.961782] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.976300] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106535, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.058166] env[65121]: WARNING neutronclient.v2_0.client [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.155037] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106530, 'name': CloneVM_Task, 'duration_secs': 1.923015} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.155320] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Created linked-clone VM from snapshot [ 838.156136] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887cea68-d4e8-4a1e-acac-e8ba8cfccb1c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.164804] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Uploading image e193384d-79f8-40cc-874e-60556f28628f {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 838.195730] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 838.195730] env[65121]: value = "vm-993434" [ 838.195730] env[65121]: _type = "VirtualMachine" [ 838.195730] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 838.196134] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-085f178e-7929-4295-9b63-31d36f1cfd08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.204565] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease: (returnval){ [ 838.204565] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521933d7-dc68-431f-846e-e3a5a5c209c2" [ 838.204565] env[65121]: _type = "HttpNfcLease" [ 838.204565] env[65121]: } obtained for exporting VM: (result){ [ 838.204565] env[65121]: value = "vm-993434" [ 838.204565] env[65121]: _type = "VirtualMachine" [ 838.204565] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 838.204978] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the lease: (returnval){ [ 838.204978] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521933d7-dc68-431f-846e-e3a5a5c209c2" [ 838.204978] env[65121]: _type = "HttpNfcLease" [ 838.204978] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 838.212419] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 838.212419] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521933d7-dc68-431f-846e-e3a5a5c209c2" [ 838.212419] env[65121]: _type = "HttpNfcLease" [ 838.212419] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 838.352547] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106532, 'name': ReconfigVM_Task, 'duration_secs': 0.663006} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.352920] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Reconfigured VM instance instance-00000039 to attach disk [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.353586] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73618c64-5411-4be2-8198-bb5cec40b792 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.362500] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 838.362500] env[65121]: value = "task-5106537" [ 838.362500] env[65121]: _type = "Task" [ 838.362500] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.375058] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106537, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.405920] env[65121]: WARNING neutronclient.v2_0.client [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.406692] env[65121]: WARNING openstack [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.407893] env[65121]: WARNING openstack [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.428664] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.428875] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.430146] env[65121]: DEBUG nova.network.neutron [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 838.485435] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106535, 'name': CreateVM_Task, 'duration_secs': 0.45577} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.489094] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.489956] env[65121]: WARNING neutronclient.v2_0.client [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.489956] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.489956] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.490277] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 838.490545] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cadf560-9afa-4804-940f-2f864c728de2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.497594] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 838.497594] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525306c2-aa4b-24b3-047d-41bbb0c3c66a" [ 838.497594] env[65121]: _type = "Task" [ 838.497594] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.510915] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525306c2-aa4b-24b3-047d-41bbb0c3c66a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.715240] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 838.715240] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521933d7-dc68-431f-846e-e3a5a5c209c2" [ 838.715240] env[65121]: _type = "HttpNfcLease" [ 838.715240] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 838.715592] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 838.715592] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521933d7-dc68-431f-846e-e3a5a5c209c2" [ 838.715592] env[65121]: _type = "HttpNfcLease" [ 838.715592] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 838.716393] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081ed08b-10a8-4fcd-8ab5-e6a2ed0c93c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.726273] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d99ba-1700-4ba7-cd61-12a39f78f367/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 838.726476] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d99ba-1700-4ba7-cd61-12a39f78f367/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 838.864331] env[65121]: WARNING openstack [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.864973] env[65121]: WARNING openstack [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.883694] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106537, 'name': Rename_Task, 'duration_secs': 0.195951} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.886594] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.887628] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5adb3d7-7fd6-4700-817b-5910b27965e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.900650] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 838.900650] env[65121]: value = "task-5106538" [ 838.900650] env[65121]: _type = "Task" [ 838.900650] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.923458] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.926387] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-651cb955-96b9-4cc2-ac5f-86c1a12689a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.932509] env[65121]: WARNING neutronclient.v2_0.client [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.933269] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.933745] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.979702] env[65121]: WARNING neutronclient.v2_0.client [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.980493] env[65121]: WARNING openstack [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.980863] env[65121]: WARNING openstack [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.012465] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525306c2-aa4b-24b3-047d-41bbb0c3c66a, 'name': SearchDatastore_Task, 'duration_secs': 0.016355} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.012465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.012465] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.012465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.012681] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.012717] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.012956] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ba8fdda-025e-4e64-ba49-e2fb4f3aea29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.026074] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7347b7b1-83d1-4584-b9ba-464d0711d0c5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.029438] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.029617] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.030902] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc961da8-1111-4fed-a33c-223ebfddd5e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.036857] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d36e3e-0580-4385-9c3b-104a6ba0db85 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.042314] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 839.042314] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523bf99d-615b-57dc-63f4-fe3f641970bc" [ 839.042314] env[65121]: _type = "Task" [ 839.042314] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.079459] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0236e482-e673-43fb-8a46-bb1bcc620c36 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.086961] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523bf99d-615b-57dc-63f4-fe3f641970bc, 'name': SearchDatastore_Task, 'duration_secs': 0.018102} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.089027] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ca2cd4d-8728-47d9-b0f6-6d6fd85a3559 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.100702] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec6c562-172f-4ec0-b4da-d332a4ed3ecf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.106048] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 839.106048] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c9638d-0320-dbb7-7f79-0a08eb33a650" [ 839.106048] env[65121]: _type = "Task" [ 839.106048] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.118290] env[65121]: DEBUG nova.compute.provider_tree [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 839.126069] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c9638d-0320-dbb7-7f79-0a08eb33a650, 'name': SearchDatastore_Task, 'duration_secs': 0.012433} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.126348] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.126662] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.126871] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-469c87ab-e370-4ebb-bca1-bc1079c674d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.135328] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 839.135328] env[65121]: value = "task-5106539" [ 839.135328] env[65121]: _type = "Task" [ 839.135328] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.147311] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106539, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.207027] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb7fc5-3096-8d91-ded2-2af022b52965/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 839.207566] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8564aea4-070d-4d93-8275-24f647c7a7cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.216526] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb7fc5-3096-8d91-ded2-2af022b52965/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 839.216823] env[65121]: ERROR oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb7fc5-3096-8d91-ded2-2af022b52965/disk-0.vmdk due to incomplete transfer. [ 839.217181] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6d0adf2d-99ac-4664-a772-fd3362b87069 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.243913] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb7fc5-3096-8d91-ded2-2af022b52965/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 839.244144] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Uploaded image 4420ebad-90bb-4083-ad61-fa47fb6880b2 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 839.247097] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 839.247483] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aff8c8cc-8260-4f90-a34d-49b4f29d6e54 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.256246] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 839.256246] env[65121]: value = "task-5106540" [ 839.256246] env[65121]: _type = "Task" [ 839.256246] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.266365] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106540, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.363822] env[65121]: DEBUG nova.network.neutron [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updated VIF entry in instance network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 839.364382] env[65121]: DEBUG nova.network.neutron [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e0942f-50", "ovs_interfaceid": "71e0942f-5026-4128-ba81-16311feb9b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 839.413262] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106538, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.425596] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.425909] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.530531] env[65121]: WARNING neutronclient.v2_0.client [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 839.531286] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.531838] env[65121]: WARNING openstack [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.653828] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106539, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.675515] env[65121]: DEBUG nova.network.neutron [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance_info_cache with network_info: [{"id": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "address": "fa:16:3e:c8:d8:39", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79bfb570-6e", "ovs_interfaceid": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 839.689099] env[65121]: DEBUG nova.scheduler.client.report [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 87 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 839.689946] env[65121]: DEBUG nova.compute.provider_tree [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 87 to 88 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 839.692817] env[65121]: DEBUG nova.compute.provider_tree [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 839.770808] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106540, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.867762] env[65121]: DEBUG oslo_concurrency.lockutils [req-d9ca2a17-939a-4539-b2c3-56bf5847d8a9 req-ab7564d8-26e1-4823-b6e7-cd8fea24fd8b service nova] Releasing lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.917132] env[65121]: DEBUG oslo_vmware.api [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106538, 'name': PowerOnVM_Task, 'duration_secs': 0.798765} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.918077] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.918659] env[65121]: INFO nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Took 7.42 seconds to spawn the instance on the hypervisor. [ 839.919053] env[65121]: DEBUG nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 839.920074] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7d7d90-79d5-4b77-88e1-ab8333baee1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.148585] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106539, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675122} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.149028] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.149131] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.149333] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdb232f2-dc90-46b0-8e3b-565890affc4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.156792] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 840.156792] env[65121]: value = "task-5106541" [ 840.156792] env[65121]: _type = "Task" [ 840.156792] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.166328] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.181197] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 840.203528] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.204057] env[65121]: DEBUG nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 840.206936] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.897s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 840.207205] env[65121]: DEBUG nova.objects.instance [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lazy-loading 'resources' on Instance uuid 00362477-c89e-4f60-98a4-d4928081d55e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.266732] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106540, 'name': Destroy_Task, 'duration_secs': 0.775785} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.267124] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Destroyed the VM [ 840.267745] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 840.267745] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9dd73878-6f7b-4cf6-8957-1666470a26b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.277124] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 840.277124] env[65121]: value = "task-5106542" [ 840.277124] env[65121]: _type = "Task" [ 840.277124] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.290059] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106542, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.441441] env[65121]: INFO nova.compute.manager [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Took 52.57 seconds to build instance. [ 840.668149] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069061} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.668427] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.669234] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51e7285-d476-4b44-a0cc-f78fb7da03b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.700449] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.703233] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10d21d66-0b92-4b61-be3a-7adf4b2aab5b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.719444] env[65121]: DEBUG nova.compute.utils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 840.725161] env[65121]: DEBUG nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 840.725422] env[65121]: DEBUG nova.network.neutron [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 840.725877] env[65121]: WARNING neutronclient.v2_0.client [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.726751] env[65121]: WARNING neutronclient.v2_0.client [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.727047] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.727471] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.738035] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef9f238-b22b-4234-b9fc-a47c6ce7029a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.742332] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 840.742332] env[65121]: value = "task-5106543" [ 840.742332] env[65121]: _type = "Task" [ 840.742332] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.767814] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7f9c06-9f7b-4e85-a8f1-d19b17937a7d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.777282] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.785903] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance '293f93f2-c01d-42c8-b1a7-3056805c77de' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 840.796618] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106542, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.814188] env[65121]: DEBUG nova.policy [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c865e09a2a0745948db257155fe8a188', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '741e03af2f4d451d9ed77004b644deb8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 840.943866] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c5a35a3d-1833-4fbf-9277-ed954375b914 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.672s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.163982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "2e676b64-181e-4fee-8120-05cb49b5bb4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.164272] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "2e676b64-181e-4fee-8120-05cb49b5bb4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.230164] env[65121]: DEBUG nova.compute.utils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 841.255742] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106543, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.273505] env[65121]: INFO nova.compute.manager [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Rebuilding instance [ 841.288854] env[65121]: DEBUG nova.network.neutron [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Successfully created port: 2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 841.293309] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.296929] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eebfc9ac-8d49-4724-b1be-b36d516d7598 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.298725] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106542, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.312606] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 841.312606] env[65121]: value = "task-5106544" [ 841.312606] env[65121]: _type = "Task" [ 841.312606] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.325040] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106544, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.333475] env[65121]: DEBUG nova.compute.manager [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 841.334441] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b65d70-4216-404f-a21c-c9454f59dae4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.429290] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0f676a-f6a0-4857-81f2-5a356dca09a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.439832] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136306e0-da45-4daf-9689-889e280dc829 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.448021] env[65121]: DEBUG nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 841.481699] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc67c3e7-7976-47a9-859d-4cc91313ac24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.490584] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d978186-930a-41fb-845f-8a0c63b5ffc0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.507711] env[65121]: DEBUG nova.compute.provider_tree [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.733401] env[65121]: DEBUG nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 841.754395] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106543, 'name': ReconfigVM_Task, 'duration_secs': 0.530982} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.754978] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfigured VM instance instance-0000003a to attach disk [datastore2] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.755493] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ef2e41a-62cc-4edd-b81e-f0a7fe40932f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.764931] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 841.764931] env[65121]: value = "task-5106545" [ 841.764931] env[65121]: _type = "Task" [ 841.764931] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.775561] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106545, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.790793] env[65121]: DEBUG oslo_vmware.api [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106542, 'name': RemoveSnapshot_Task, 'duration_secs': 1.104778} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.792686] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 841.793492] env[65121]: INFO nova.compute.manager [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Took 19.06 seconds to snapshot the instance on the hypervisor. [ 841.823527] env[65121]: DEBUG oslo_vmware.api [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106544, 'name': PowerOnVM_Task, 'duration_secs': 0.446708} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.824227] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.824227] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e9cfc6-dce4-4785-9f22-26425353a3d4 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance '293f93f2-c01d-42c8-b1a7-3056805c77de' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 842.012864] env[65121]: DEBUG nova.scheduler.client.report [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 842.082290] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.279482] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106545, 'name': Rename_Task, 'duration_secs': 0.317797} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.279805] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.280130] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e49f4c3-be6c-4528-8aba-19325b18deb4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.290737] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 842.290737] env[65121]: value = "task-5106546" [ 842.290737] env[65121]: _type = "Task" [ 842.290737] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.306644] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106546, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.357205] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.357679] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c14674b9-d6dd-4a0b-98e9-09b7a869dd9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.368311] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 842.368311] env[65121]: value = "task-5106547" [ 842.368311] env[65121]: _type = "Task" [ 842.368311] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.376042] env[65121]: DEBUG nova.compute.manager [None req-6de8e67f-4c16-4dd0-88d2-3b7bb38d564d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Found 2 images (rotation: 2) {{(pid=65121) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 842.381363] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.519282] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.312s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.522763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 40.617s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.522763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.522763] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 842.523067] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.137s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.523305] env[65121]: DEBUG nova.objects.instance [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lazy-loading 'resources' on Instance uuid 19b1f3f9-842e-4150-8890-b0b22393c3af {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.525824] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918c741a-f0a7-49d9-a866-8ba38fcc3432 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.536378] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f1ea2f-6fd2-445c-8a68-6604da8269b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.555379] env[65121]: INFO nova.scheduler.client.report [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Deleted allocations for instance 00362477-c89e-4f60-98a4-d4928081d55e [ 842.557217] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75593fd-b5ce-4996-aa25-46fbf15955e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.567612] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ac3c97-e17b-42f8-b3cb-f3440ba539a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.602294] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177269MB free_disk=94GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 842.602447] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.745769] env[65121]: DEBUG nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 842.803730] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106546, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.880539] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106547, 'name': PowerOffVM_Task, 'duration_secs': 0.149877} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.880862] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 842.881167] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 842.883024] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5e59a8-8d34-491a-a797-aec740add742 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.890810] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 842.891140] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2f0876e-24ad-426b-987b-b3c03995bb3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.901413] env[65121]: DEBUG nova.network.neutron [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Successfully updated port: 2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 842.920016] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 842.920284] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 842.920441] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Deleting the datastore file [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.920723] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aac07542-11e7-47bb-855b-ddd5ad5a8b5b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.930136] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 842.930136] env[65121]: value = "task-5106549" [ 842.930136] env[65121]: _type = "Task" [ 842.930136] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.934615] env[65121]: DEBUG nova.compute.manager [req-633d0000-69e0-43c2-883d-7c56461065f3 req-56895f2a-6423-43b4-bf67-22d22620102a service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Received event network-vif-plugged-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 842.934957] env[65121]: DEBUG oslo_concurrency.lockutils [req-633d0000-69e0-43c2-883d-7c56461065f3 req-56895f2a-6423-43b4-bf67-22d22620102a service nova] Acquiring lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.936131] env[65121]: DEBUG oslo_concurrency.lockutils [req-633d0000-69e0-43c2-883d-7c56461065f3 req-56895f2a-6423-43b4-bf67-22d22620102a service nova] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.936131] env[65121]: DEBUG oslo_concurrency.lockutils [req-633d0000-69e0-43c2-883d-7c56461065f3 req-56895f2a-6423-43b4-bf67-22d22620102a service nova] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.936131] env[65121]: DEBUG nova.compute.manager [req-633d0000-69e0-43c2-883d-7c56461065f3 req-56895f2a-6423-43b4-bf67-22d22620102a service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] No waiting events found dispatching network-vif-plugged-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 842.936131] env[65121]: WARNING nova.compute.manager [req-633d0000-69e0-43c2-883d-7c56461065f3 req-56895f2a-6423-43b4-bf67-22d22620102a service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Received unexpected event network-vif-plugged-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c for instance with vm_state building and task_state spawning. [ 842.947642] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.069546] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49a6ed8b-807a-428d-b704-244e75615764 tempest-ServerPasswordTestJSON-1617341377 tempest-ServerPasswordTestJSON-1617341377-project-member] Lock "00362477-c89e-4f60-98a4-d4928081d55e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.564s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.082507] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:20:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='1508158469',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-1961943058',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 843.082749] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 843.082898] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 843.083122] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 843.083311] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 843.083490] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 843.083705] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.083861] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 843.086027] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 843.086435] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 843.086583] env[65121]: DEBUG nova.virt.hardware [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 843.087545] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa91813b-06c8-443e-afd1-3972f7290310 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.109518] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f2baa4-2c5d-4d56-9cf9-2e3e4ac5e74a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.300772] env[65121]: DEBUG oslo_vmware.api [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106546, 'name': PowerOnVM_Task, 'duration_secs': 0.709727} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.301121] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.301297] env[65121]: INFO nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Took 7.86 seconds to spawn the instance on the hypervisor. [ 843.301460] env[65121]: DEBUG nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 843.304928] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca464cf6-daa0-4306-a747-f94771e19540 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.404789] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.405747] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.405747] env[65121]: DEBUG nova.network.neutron [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 843.441264] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207819} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.445214] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.445214] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.445214] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.608410] env[65121]: DEBUG nova.compute.manager [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 843.609519] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92465135-2448-48f0-a02c-340e8c13d201 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.684870] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2819b773-c779-417a-95cd-ca1256c3e1c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.693819] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b38d93a-83d8-44ea-8366-c05d38732978 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.733293] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3172bc-6455-4c03-a68d-2cf9abbc9089 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.742384] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b5028f-bd6f-4ae6-adad-7c689c714e3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.757816] env[65121]: DEBUG nova.compute.provider_tree [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.828543] env[65121]: INFO nova.compute.manager [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Took 55.36 seconds to build instance. [ 843.908164] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.908620] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.954145] env[65121]: DEBUG nova.network.neutron [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 844.014620] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.017918] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.124138] env[65121]: INFO nova.compute.manager [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] instance snapshotting [ 844.124807] env[65121]: DEBUG nova.objects.instance [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'flavor' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.261290] env[65121]: DEBUG nova.scheduler.client.report [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 844.332749] env[65121]: DEBUG oslo_concurrency.lockutils [None req-be90a1dd-7b26-4284-9eb2-58df5133998d tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.972s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.486492] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 844.487159] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 844.487466] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 844.487925] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 844.488351] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 844.488725] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 844.489218] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.490654] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 844.490654] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 844.490654] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 844.490654] env[65121]: DEBUG nova.virt.hardware [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 844.491539] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f35487-4402-4b65-96a0-8d8c73e8ac95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.501025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea39709-9f3a-461c-871c-50289c938774 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.517508] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.523453] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 844.523740] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.523961] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a0e633b-b3be-4ff7-afb2-60fbf21c9508 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.551287] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.551287] env[65121]: value = "task-5106550" [ 844.551287] env[65121]: _type = "Task" [ 844.551287] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.561852] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106550, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.573533] env[65121]: WARNING neutronclient.v2_0.client [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 844.574256] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.574617] env[65121]: WARNING openstack [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.633420] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852713b7-33fe-474a-ab2f-8cc3a1463a4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.662731] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412fdae0-6f9f-4dd4-815d-fe1eec87a031 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.726241] env[65121]: DEBUG nova.network.neutron [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Updating instance_info_cache with network_info: [{"id": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "address": "fa:16:3e:80:39:87", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0f93ed-17", "ovs_interfaceid": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 844.770709] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.247s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.778360] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.805s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.778360] env[65121]: INFO nova.compute.claims [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.808188] env[65121]: INFO nova.scheduler.client.report [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Deleted allocations for instance 19b1f3f9-842e-4150-8890-b0b22393c3af [ 844.839515] env[65121]: DEBUG nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 845.063942] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106550, 'name': CreateVM_Task, 'duration_secs': 0.413152} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.065092] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.065092] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.065092] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 845.066832] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 845.066832] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f38d04f-aac5-4ca3-8289-5b583eab6218 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.077266] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 845.077266] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ce9afd-9afc-632a-dc52-f899f73056c1" [ 845.077266] env[65121]: _type = "Task" [ 845.077266] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.087367] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ce9afd-9afc-632a-dc52-f899f73056c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.175814] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 845.176228] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c35c0460-f585-4314-8f70-28e97084830f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.188693] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 845.188693] env[65121]: value = "task-5106551" [ 845.188693] env[65121]: _type = "Task" [ 845.188693] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.198746] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106551, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.229672] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Releasing lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 845.230055] env[65121]: DEBUG nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Instance network_info: |[{"id": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "address": "fa:16:3e:80:39:87", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0f93ed-17", "ovs_interfaceid": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 845.230560] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:39:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e0f93ed-17fe-4ad2-aadd-13b7bdac954c', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.242301] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 845.242301] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.242509] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7aaa426-0449-4e36-93cc-956bb7a80ac1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.268733] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.268733] env[65121]: value = "task-5106552" [ 845.268733] env[65121]: _type = "Task" [ 845.268733] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.284800] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106552, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.287154] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.317547] env[65121]: DEBUG oslo_concurrency.lockutils [None req-089ce22c-4d08-474b-9425-8485f7312b4e tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "19b1f3f9-842e-4150-8890-b0b22393c3af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.044s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.371877] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.395878] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.396530] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.458974] env[65121]: DEBUG nova.network.neutron [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Port 79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 845.459328] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.460774] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 845.460774] env[65121]: DEBUG nova.network.neutron [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 845.549409] env[65121]: DEBUG nova.compute.manager [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Received event network-changed-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 845.550836] env[65121]: DEBUG nova.compute.manager [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Refreshing instance network info cache due to event network-changed-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 845.551217] env[65121]: DEBUG oslo_concurrency.lockutils [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Acquiring lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.551442] env[65121]: DEBUG oslo_concurrency.lockutils [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Acquired lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 845.551635] env[65121]: DEBUG nova.network.neutron [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Refreshing network info cache for port 2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 845.589252] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ce9afd-9afc-632a-dc52-f899f73056c1, 'name': SearchDatastore_Task, 'duration_secs': 0.026205} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.589600] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 845.589861] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.590123] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.590285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 845.590499] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.590782] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78136754-6e11-4614-a6a3-33a88519d21a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.603400] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.603400] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.604295] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7786bd1d-618b-496a-9af3-90683bc2d464 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.614198] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 845.614198] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529d04b1-95c9-985e-6052-1bb80aecbdcd" [ 845.614198] env[65121]: _type = "Task" [ 845.614198] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.625659] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529d04b1-95c9-985e-6052-1bb80aecbdcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.703828] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106551, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.778151] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106552, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.965194] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.966018] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.966406] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.058921] env[65121]: WARNING neutronclient.v2_0.client [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.059951] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.060331] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.105893] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.105893] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.132509] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529d04b1-95c9-985e-6052-1bb80aecbdcd, 'name': SearchDatastore_Task, 'duration_secs': 0.019489} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.133353] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f1d74e8-e5c7-472f-bf42-02d059a376be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.142284] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 846.142284] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cb33fa-2d68-15b2-cf77-7654a95077f9" [ 846.142284] env[65121]: _type = "Task" [ 846.142284] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.152214] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cb33fa-2d68-15b2-cf77-7654a95077f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.203737] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106551, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.207582] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.208239] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.208600] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.217392] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.217774] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.283177] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106552, 'name': CreateVM_Task, 'duration_secs': 0.52421} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.283422] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.284103] env[65121]: WARNING neutronclient.v2_0.client [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.284558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.284879] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.285291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 846.285795] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68b22d73-f4cc-4abf-b6e9-742b3f46a903 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.298469] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 846.298469] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a23235-1da6-6b3c-007d-d5e5afd8183d" [ 846.298469] env[65121]: _type = "Task" [ 846.298469] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.310852] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a23235-1da6-6b3c-007d-d5e5afd8183d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.339684] env[65121]: WARNING neutronclient.v2_0.client [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.340403] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.340717] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.352839] env[65121]: DEBUG nova.network.neutron [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance_info_cache with network_info: [{"id": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "address": "fa:16:3e:c8:d8:39", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79bfb570-6e", "ovs_interfaceid": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 846.429412] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90a98fc-705a-4b6f-a95d-96bb40f6ee72 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.440321] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6c910e-61c2-485d-8037-f4105bd25678 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.473835] env[65121]: DEBUG nova.network.neutron [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Updated VIF entry in instance network info cache for port 2e0f93ed-17fe-4ad2-aadd-13b7bdac954c. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 846.473835] env[65121]: DEBUG nova.network.neutron [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Updating instance_info_cache with network_info: [{"id": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "address": "fa:16:3e:80:39:87", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0f93ed-17", "ovs_interfaceid": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 846.475568] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d46d83-3eef-4948-933d-53584a771e7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.484765] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682cd1eb-01ed-44fd-97b7-5ec819dcea3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.502027] env[65121]: DEBUG nova.compute.provider_tree [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.654329] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cb33fa-2d68-15b2-cf77-7654a95077f9, 'name': SearchDatastore_Task, 'duration_secs': 0.01699} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.654640] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.654872] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.655304] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f3730a9-fc41-4bef-9929-9eb984d7409d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.661859] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 846.661859] env[65121]: value = "task-5106553" [ 846.661859] env[65121]: _type = "Task" [ 846.661859] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.670591] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.700972] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106551, 'name': CreateSnapshot_Task, 'duration_secs': 1.259792} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.701265] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 846.702037] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8fb440-b567-44eb-8ccf-81926f62f2b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.810569] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a23235-1da6-6b3c-007d-d5e5afd8183d, 'name': SearchDatastore_Task, 'duration_secs': 0.022974} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.810911] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.811534] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.811534] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.811534] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.812045] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.812045] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55727383-5eff-409d-ab8b-8115302ebd53 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.825409] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.825644] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.826286] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2c9fe1b-2124-4397-af40-ef4778689a2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.833064] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 846.833064] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52378dbd-4769-772b-c774-6f104103468c" [ 846.833064] env[65121]: _type = "Task" [ 846.833064] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.843443] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52378dbd-4769-772b-c774-6f104103468c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.859727] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.979876] env[65121]: DEBUG oslo_concurrency.lockutils [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Releasing lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.980269] env[65121]: DEBUG nova.compute.manager [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 846.980339] env[65121]: DEBUG nova.compute.manager [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing instance network info cache due to event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 846.980523] env[65121]: DEBUG oslo_concurrency.lockutils [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Acquiring lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.980711] env[65121]: DEBUG oslo_concurrency.lockutils [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Acquired lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.980791] env[65121]: DEBUG nova.network.neutron [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 847.004984] env[65121]: DEBUG nova.scheduler.client.report [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 847.172854] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106553, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.222050] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 847.222460] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5057fac8-4c40-47a2-89f0-1f68a27cb5e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.231710] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 847.231710] env[65121]: value = "task-5106554" [ 847.231710] env[65121]: _type = "Task" [ 847.231710] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.243550] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106554, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.346451] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52378dbd-4769-772b-c774-6f104103468c, 'name': SearchDatastore_Task, 'duration_secs': 0.027772} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.347313] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6022d59-0ed6-4633-a534-1885ed7e960c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.357457] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 847.357457] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52915808-eb12-3bcf-23bc-ca9be244b786" [ 847.357457] env[65121]: _type = "Task" [ 847.357457] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.364123] env[65121]: DEBUG nova.compute.manager [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65121) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 847.364639] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 847.369906] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52915808-eb12-3bcf-23bc-ca9be244b786, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.484196] env[65121]: WARNING neutronclient.v2_0.client [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 847.484902] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.485444] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.510651] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.736s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.511069] env[65121]: DEBUG nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 847.517081] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.442s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.518669] env[65121]: INFO nova.compute.claims [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.682651] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106553, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.967175} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.682810] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.683051] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.694795] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.694795] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.694795] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a646ed5b-d3ef-4799-99b9-fd0bf6750263 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.701576] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 847.701576] env[65121]: value = "task-5106555" [ 847.701576] env[65121]: _type = "Task" [ 847.701576] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.715375] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.742748] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106554, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.801210] env[65121]: WARNING neutronclient.v2_0.client [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 847.801429] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.801772] env[65121]: WARNING openstack [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.871613] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52915808-eb12-3bcf-23bc-ca9be244b786, 'name': SearchDatastore_Task, 'duration_secs': 0.075901} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.871880] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 847.871918] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] b22ccce7-c54a-4577-9de0-1fd9c10cd189/b22ccce7-c54a-4577-9de0-1fd9c10cd189.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.872243] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2d47287-7265-429a-8524-4a8342d6dc32 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.882899] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 847.882899] env[65121]: value = "task-5106556" [ 847.882899] env[65121]: _type = "Task" [ 847.882899] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.893389] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.916330] env[65121]: DEBUG nova.network.neutron [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updated VIF entry in instance network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 847.916990] env[65121]: DEBUG nova.network.neutron [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e0942f-50", "ovs_interfaceid": "71e0942f-5026-4128-ba81-16311feb9b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 847.976284] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "600d5554-f52a-48ca-941d-1a755d086823" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 847.976513] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 848.027768] env[65121]: DEBUG nova.compute.utils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 848.029211] env[65121]: DEBUG nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 848.029587] env[65121]: DEBUG nova.network.neutron [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 848.031064] env[65121]: WARNING neutronclient.v2_0.client [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 848.031064] env[65121]: WARNING neutronclient.v2_0.client [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 848.031064] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 848.031239] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 848.041732] env[65121]: DEBUG nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 848.088294] env[65121]: DEBUG nova.policy [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1f089b8cd61d42e196f268426a2a59c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7d185e5679745caa16ec007c348797b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 848.212993] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.15116} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.213164] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.214497] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9f9738-d285-494b-8f2e-3583252989a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.239557] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.240769] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be5d5cac-95a5-4571-bde5-e52c624bc732 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.272744] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106554, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.275233] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 848.275233] env[65121]: value = "task-5106557" [ 848.275233] env[65121]: _type = "Task" [ 848.275233] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.285568] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106557, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.398017] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106556, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.419780] env[65121]: DEBUG oslo_concurrency.lockutils [req-8c1590e3-b3ed-40c1-a3e4-2eade08371bd req-8bb1c935-6272-4b85-a0b5-a53dcdff912c service nova] Releasing lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.753676] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106554, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.790277] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106557, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.898486] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685617} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.898823] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] b22ccce7-c54a-4577-9de0-1fd9c10cd189/b22ccce7-c54a-4577-9de0-1fd9c10cd189.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.898968] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.899266] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c75f9814-4542-4dc2-af51-9fadc9cfe950 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.910850] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 848.910850] env[65121]: value = "task-5106558" [ 848.910850] env[65121]: _type = "Task" [ 848.910850] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.926421] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106558, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.959673] env[65121]: DEBUG nova.network.neutron [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Successfully created port: e904441b-001c-4583-a1e4-98b5fbd6eb74 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 849.055985] env[65121]: DEBUG nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 849.088729] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 849.089081] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 849.089081] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 849.089272] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 849.089481] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 849.089726] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 849.090015] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.090186] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 849.090362] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 849.090529] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 849.090699] env[65121]: DEBUG nova.virt.hardware [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 849.092267] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae8a566-e8ee-4813-ac0e-596c2babce9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.104406] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24df54e-02be-4f35-b8d3-30ada5135028 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.197457] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70f7262-7bdd-446f-b7ae-7ddc1a3962d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.207283] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80496bcd-ed45-44ff-bbbb-b23c8778923c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.247728] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9659ef-164d-4185-8468-b0439f5662bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.257405] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106554, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.261836] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97eecb2-e30e-4f90-b40c-33322bced70c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.279626] env[65121]: DEBUG nova.compute.provider_tree [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.290957] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106557, 'name': ReconfigVM_Task, 'duration_secs': 1.004898} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.292579] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Reconfigured VM instance instance-00000039 to attach disk [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec/e2e03e5c-4a71-4555-9cd4-d67e178064ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.292579] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-176bd05c-acb2-460a-9298-73a617c279be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.300041] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 849.300041] env[65121]: value = "task-5106559" [ 849.300041] env[65121]: _type = "Task" [ 849.300041] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.312117] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106559, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.421446] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106558, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157737} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.421714] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.422548] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce8e084-d8c9-4c86-89fc-206d987b14f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.447586] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] b22ccce7-c54a-4577-9de0-1fd9c10cd189/b22ccce7-c54a-4577-9de0-1fd9c10cd189.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.447586] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cc0ea6a-3351-4215-8e1a-0089bafa2010 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.467707] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 849.467707] env[65121]: value = "task-5106560" [ 849.467707] env[65121]: _type = "Task" [ 849.467707] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.476837] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106560, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.575767] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d99ba-1700-4ba7-cd61-12a39f78f367/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 849.578688] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65ded15-7217-4091-8efb-df6f95521d56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.591417] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d99ba-1700-4ba7-cd61-12a39f78f367/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 849.591737] env[65121]: ERROR oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d99ba-1700-4ba7-cd61-12a39f78f367/disk-0.vmdk due to incomplete transfer. [ 849.592121] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-34c46d4d-9315-4b09-98e4-f8c12ac969ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.601119] env[65121]: DEBUG oslo_vmware.rw_handles [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d99ba-1700-4ba7-cd61-12a39f78f367/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 849.601328] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Uploaded image e193384d-79f8-40cc-874e-60556f28628f to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 849.603235] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 849.603490] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1b017db8-372e-4ede-a1b6-7c3574094c34 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.610387] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 849.610387] env[65121]: value = "task-5106561" [ 849.610387] env[65121]: _type = "Task" [ 849.610387] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.620578] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106561, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.636385] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "f1a5fdce-a940-43d4-83d7-716786c9ac34" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.636705] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.636932] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "f1a5fdce-a940-43d4-83d7-716786c9ac34-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.637148] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.637352] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.640026] env[65121]: INFO nova.compute.manager [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Terminating instance [ 849.753176] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106554, 'name': CloneVM_Task, 'duration_secs': 2.329054} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.753475] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Created linked-clone VM from snapshot [ 849.754356] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fc8860-01b9-4e71-94bf-7045ecaa4ecd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.762606] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Uploading image 6b971b17-fd69-4b46-8211-01caca537cc1 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 849.782181] env[65121]: DEBUG nova.scheduler.client.report [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 849.794635] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 849.794635] env[65121]: value = "vm-993441" [ 849.794635] env[65121]: _type = "VirtualMachine" [ 849.794635] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 849.794635] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a2948470-2442-4018-97aa-c5bee2b240e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.810042] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease: (returnval){ [ 849.810042] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5243df59-5b7e-b258-f611-64e912809f65" [ 849.810042] env[65121]: _type = "HttpNfcLease" [ 849.810042] env[65121]: } obtained for exporting VM: (result){ [ 849.810042] env[65121]: value = "vm-993441" [ 849.810042] env[65121]: _type = "VirtualMachine" [ 849.810042] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 849.810042] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the lease: (returnval){ [ 849.810042] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5243df59-5b7e-b258-f611-64e912809f65" [ 849.810042] env[65121]: _type = "HttpNfcLease" [ 849.810042] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 849.821252] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106559, 'name': Rename_Task, 'duration_secs': 0.317448} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.822236] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.822603] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9f0d384-efde-43df-af94-a9f3b0094231 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.827375] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.827375] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5243df59-5b7e-b258-f611-64e912809f65" [ 849.827375] env[65121]: _type = "HttpNfcLease" [ 849.827375] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 849.828173] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 849.828173] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5243df59-5b7e-b258-f611-64e912809f65" [ 849.828173] env[65121]: _type = "HttpNfcLease" [ 849.828173] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 849.829254] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe0c00d-a506-4a83-a33b-411f18de9511 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.833886] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 849.833886] env[65121]: value = "task-5106563" [ 849.833886] env[65121]: _type = "Task" [ 849.833886] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.840461] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5202f174-c044-b118-21f1-13615df5ac06/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 849.840861] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5202f174-c044-b118-21f1-13615df5ac06/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 849.899116] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106563, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.982753] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106560, 'name': ReconfigVM_Task, 'duration_secs': 0.314243} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.983060] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Reconfigured VM instance instance-0000003b to attach disk [datastore2] b22ccce7-c54a-4577-9de0-1fd9c10cd189/b22ccce7-c54a-4577-9de0-1fd9c10cd189.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.983365] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=65121) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 849.984073] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-1db6e6fe-506a-4401-8182-8757301a57b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.992572] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 849.992572] env[65121]: value = "task-5106564" [ 849.992572] env[65121]: _type = "Task" [ 849.992572] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.003395] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106564, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.042570] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-068be17d-26af-4795-a820-cc8abf130bf0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.124209] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106561, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.144246] env[65121]: DEBUG nova.compute.manager [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 850.144483] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.146477] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330325c0-8cea-40dc-9013-a363ad6e128e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.154616] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.154616] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd0094c6-b38e-4e1f-8135-03496b13fa92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.162492] env[65121]: DEBUG oslo_vmware.api [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 850.162492] env[65121]: value = "task-5106565" [ 850.162492] env[65121]: _type = "Task" [ 850.162492] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.175296] env[65121]: DEBUG oslo_vmware.api [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.288140] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.771s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.288727] env[65121]: DEBUG nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 850.291633] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.632s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.291853] env[65121]: DEBUG nova.objects.instance [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lazy-loading 'resources' on Instance uuid c1632a96-f334-4978-9f31-97f9329577e5 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.347715] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106563, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.503333] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106564, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.064094} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.503744] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=65121) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 850.504402] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4674c65-1f4e-41e3-b6ce-aacbac10e30e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.531712] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] b22ccce7-c54a-4577-9de0-1fd9c10cd189/ephemeral_0.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.532164] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a8b2ccf-9cc0-4f21-a9df-819345d7051e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.553050] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 850.553050] env[65121]: value = "task-5106566" [ 850.553050] env[65121]: _type = "Task" [ 850.553050] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.564091] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106566, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.622351] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106561, 'name': Destroy_Task, 'duration_secs': 0.799388} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.623687] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Destroyed the VM [ 850.623687] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 850.624017] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1a5b2c47-bb54-4da5-ad35-238d4c017aa4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.633041] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 850.633041] env[65121]: value = "task-5106567" [ 850.633041] env[65121]: _type = "Task" [ 850.633041] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.640371] env[65121]: DEBUG nova.network.neutron [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Successfully updated port: e904441b-001c-4583-a1e4-98b5fbd6eb74 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 850.648920] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106567, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.681754] env[65121]: DEBUG oslo_vmware.api [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106565, 'name': PowerOffVM_Task, 'duration_secs': 0.252478} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.683815] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.683815] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.684758] env[65121]: DEBUG nova.compute.manager [req-6010f09f-a183-4a78-a5c8-bfae23b01001 req-d295c2cf-906d-4b54-9681-6a5bb1217573 service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Received event network-vif-plugged-e904441b-001c-4583-a1e4-98b5fbd6eb74 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 850.685011] env[65121]: DEBUG oslo_concurrency.lockutils [req-6010f09f-a183-4a78-a5c8-bfae23b01001 req-d295c2cf-906d-4b54-9681-6a5bb1217573 service nova] Acquiring lock "5d352184-7074-4fab-b4ac-cc11a6b936b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.685280] env[65121]: DEBUG oslo_concurrency.lockutils [req-6010f09f-a183-4a78-a5c8-bfae23b01001 req-d295c2cf-906d-4b54-9681-6a5bb1217573 service nova] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.686485] env[65121]: DEBUG oslo_concurrency.lockutils [req-6010f09f-a183-4a78-a5c8-bfae23b01001 req-d295c2cf-906d-4b54-9681-6a5bb1217573 service nova] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.686485] env[65121]: DEBUG nova.compute.manager [req-6010f09f-a183-4a78-a5c8-bfae23b01001 req-d295c2cf-906d-4b54-9681-6a5bb1217573 service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] No waiting events found dispatching network-vif-plugged-e904441b-001c-4583-a1e4-98b5fbd6eb74 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 850.686485] env[65121]: WARNING nova.compute.manager [req-6010f09f-a183-4a78-a5c8-bfae23b01001 req-d295c2cf-906d-4b54-9681-6a5bb1217573 service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Received unexpected event network-vif-plugged-e904441b-001c-4583-a1e4-98b5fbd6eb74 for instance with vm_state building and task_state spawning. [ 850.688322] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b224742-e842-4ac4-b304-2483d11f36bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.761621] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 850.761870] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 850.762022] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Deleting the datastore file [datastore1] f1a5fdce-a940-43d4-83d7-716786c9ac34 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.762330] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e1bfe8c-e8ee-4794-9b02-229197083c0f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.770928] env[65121]: DEBUG oslo_vmware.api [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for the task: (returnval){ [ 850.770928] env[65121]: value = "task-5106569" [ 850.770928] env[65121]: _type = "Task" [ 850.770928] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.782472] env[65121]: DEBUG oslo_vmware.api [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.795534] env[65121]: DEBUG nova.compute.utils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 850.801277] env[65121]: DEBUG nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 850.801640] env[65121]: DEBUG nova.network.neutron [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 850.805076] env[65121]: WARNING neutronclient.v2_0.client [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.805076] env[65121]: WARNING neutronclient.v2_0.client [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.805076] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.805076] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.850526] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106563, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.892881] env[65121]: DEBUG nova.policy [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc3c22580f4442dcae4a2b5245ede996', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '083c4f2aa7bf4167b472db6bebb9d8bd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 851.072612] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106566, 'name': ReconfigVM_Task, 'duration_secs': 0.446655} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.073708] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Reconfigured VM instance instance-0000003b to attach disk [datastore2] b22ccce7-c54a-4577-9de0-1fd9c10cd189/ephemeral_0.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.074265] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-769a5a94-216d-459a-9be2-849f0643acbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.085354] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 851.085354] env[65121]: value = "task-5106570" [ 851.085354] env[65121]: _type = "Task" [ 851.085354] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.107918] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106570, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.149107] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106567, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.151151] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "refresh_cache-5d352184-7074-4fab-b4ac-cc11a6b936b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.151151] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquired lock "refresh_cache-5d352184-7074-4fab-b4ac-cc11a6b936b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 851.152729] env[65121]: DEBUG nova.network.neutron [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 851.283138] env[65121]: DEBUG oslo_vmware.api [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Task: {'id': task-5106569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275359} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.286760] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.287054] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.287249] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.287421] env[65121]: INFO nova.compute.manager [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Took 1.14 seconds to destroy the instance on the hypervisor. [ 851.288278] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 851.292652] env[65121]: DEBUG nova.compute.manager [-] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 851.292882] env[65121]: DEBUG nova.network.neutron [-] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 851.293140] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.293946] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.295074] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.303239] env[65121]: DEBUG nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 851.344143] env[65121]: DEBUG nova.network.neutron [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Successfully created port: 027b9031-f921-43b2-8580-de8d028869c3 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 851.347420] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.356456] env[65121]: DEBUG oslo_vmware.api [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106563, 'name': PowerOnVM_Task, 'duration_secs': 1.105049} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.359833] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.360667] env[65121]: DEBUG nova.compute.manager [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 851.362791] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef03a930-ef64-497c-83ef-49356ec68c80 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.449630] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 851.449630] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 851.598710] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106570, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.652202] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650b9b04-555b-45a8-b933-02b5b1c01234 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.655486] env[65121]: DEBUG oslo_vmware.api [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106567, 'name': RemoveSnapshot_Task, 'duration_secs': 0.937428} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.657310] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.657409] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.665873] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 851.666454] env[65121]: INFO nova.compute.manager [None req-d93acbe2-49d2-447d-ae7a-84af89573092 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Took 19.19 seconds to snapshot the instance on the hypervisor. [ 851.677206] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0d0f43-518c-430b-8346-ce75b5211f83 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.716665] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84316316-eb0c-4823-b9ec-3383effe63f3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.726843] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec7b194-6edd-4913-9d93-789a47036687 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.743045] env[65121]: DEBUG nova.compute.provider_tree [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.762206] env[65121]: DEBUG nova.network.neutron [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 851.833032] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.833032] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.885429] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 851.923394] env[65121]: WARNING neutronclient.v2_0.client [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.924195] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.924551] env[65121]: WARNING openstack [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 852.041035] env[65121]: DEBUG nova.network.neutron [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Updating instance_info_cache with network_info: [{"id": "e904441b-001c-4583-a1e4-98b5fbd6eb74", "address": "fa:16:3e:5e:ee:13", "network": {"id": "4af4fcd7-26ea-487e-a8e7-90cdec609474", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-431637469-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d185e5679745caa16ec007c348797b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape904441b-00", "ovs_interfaceid": "e904441b-001c-4583-a1e4-98b5fbd6eb74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 852.100517] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106570, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.148023] env[65121]: DEBUG nova.network.neutron [-] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 852.226740] env[65121]: DEBUG nova.compute.manager [req-1d950b7e-799a-4c40-b80b-132d8dc5e540 req-c4939ed8-874b-4142-97b8-fa3a109d64d1 service nova] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Received event network-vif-deleted-3a66c1d6-9549-459d-bb1b-66781725a57f {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 852.247166] env[65121]: DEBUG nova.scheduler.client.report [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 852.314859] env[65121]: DEBUG nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 852.345553] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 852.345819] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 852.345976] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 852.346423] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 852.347030] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 852.347030] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 852.347217] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.347327] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 852.347493] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 852.347702] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 852.349563] env[65121]: DEBUG nova.virt.hardware [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 852.349563] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c45706-4645-4bf9-95ea-6aa908dae855 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.362346] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24f47d7-ff98-4354-ba3c-93706f1ded56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.543954] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Releasing lock "refresh_cache-5d352184-7074-4fab-b4ac-cc11a6b936b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 852.544035] env[65121]: DEBUG nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Instance network_info: |[{"id": "e904441b-001c-4583-a1e4-98b5fbd6eb74", "address": "fa:16:3e:5e:ee:13", "network": {"id": "4af4fcd7-26ea-487e-a8e7-90cdec609474", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-431637469-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d185e5679745caa16ec007c348797b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape904441b-00", "ovs_interfaceid": "e904441b-001c-4583-a1e4-98b5fbd6eb74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 852.544708] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:ee:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e904441b-001c-4583-a1e4-98b5fbd6eb74', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.552647] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Creating folder: Project (b7d185e5679745caa16ec007c348797b). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.553151] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63674eb9-3af5-47e4-a449-51973bedd896 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.565961] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Created folder: Project (b7d185e5679745caa16ec007c348797b) in parent group-v993268. [ 852.566222] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Creating folder: Instances. Parent ref: group-v993442. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.566508] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c97abf7-22f6-4cea-a1ab-b8633cdef660 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.578770] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Created folder: Instances in parent group-v993442. [ 852.579803] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 852.579803] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.579803] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c23dd1d-34df-451d-92f6-6b931c43a4c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.604223] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106570, 'name': Rename_Task, 'duration_secs': 1.069735} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.606032] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.606665] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.606665] env[65121]: value = "task-5106573" [ 852.606665] env[65121]: _type = "Task" [ 852.606665] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.606897] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0f96252-fdbc-4c86-bf46-76fe21e75f40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.618313] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106573, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.619891] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 852.619891] env[65121]: value = "task-5106574" [ 852.619891] env[65121]: _type = "Task" [ 852.619891] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.628728] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106574, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.653744] env[65121]: INFO nova.compute.manager [-] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Took 1.36 seconds to deallocate network for instance. [ 852.754329] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.462s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.757793] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.755s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.758816] env[65121]: DEBUG nova.objects.instance [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lazy-loading 'resources' on Instance uuid cd86f815-4b46-45b1-9a1c-ab2e28f63626 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.783750] env[65121]: INFO nova.scheduler.client.report [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Deleted allocations for instance c1632a96-f334-4978-9f31-97f9329577e5 [ 852.800397] env[65121]: DEBUG nova.compute.manager [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Received event network-changed-e904441b-001c-4583-a1e4-98b5fbd6eb74 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 852.800732] env[65121]: DEBUG nova.compute.manager [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Refreshing instance network info cache due to event network-changed-e904441b-001c-4583-a1e4-98b5fbd6eb74. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 852.801042] env[65121]: DEBUG oslo_concurrency.lockutils [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Acquiring lock "refresh_cache-5d352184-7074-4fab-b4ac-cc11a6b936b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.801772] env[65121]: DEBUG oslo_concurrency.lockutils [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Acquired lock "refresh_cache-5d352184-7074-4fab-b4ac-cc11a6b936b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.801772] env[65121]: DEBUG nova.network.neutron [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Refreshing network info cache for port e904441b-001c-4583-a1e4-98b5fbd6eb74 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 852.975777] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 852.976140] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.810049] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.810049] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.834s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.810500] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.812637] env[65121]: DEBUG nova.network.neutron [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Successfully updated port: 027b9031-f921-43b2-8580-de8d028869c3 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 853.815763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.821622] env[65121]: WARNING neutronclient.v2_0.client [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 853.822362] env[65121]: WARNING openstack [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.822715] env[65121]: WARNING openstack [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 853.830576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b54f6b42-ee9b-4dc2-890a-715d2ab4fdb4 tempest-ServersTestMultiNic-1259526584 tempest-ServersTestMultiNic-1259526584-project-member] Lock "c1632a96-f334-4978-9f31-97f9329577e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.275s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 853.832565] env[65121]: INFO nova.compute.manager [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Terminating instance [ 853.849597] env[65121]: DEBUG oslo_vmware.api [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106574, 'name': PowerOnVM_Task, 'duration_secs': 0.536479} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.855236] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.855236] env[65121]: INFO nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Took 11.11 seconds to spawn the instance on the hypervisor. [ 853.855236] env[65121]: DEBUG nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 853.855236] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106573, 'name': CreateVM_Task, 'duration_secs': 0.431939} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.856251] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a66b7a-917f-4c1b-82ef-ec0f2c99c8bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.859262] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.859799] env[65121]: WARNING neutronclient.v2_0.client [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 853.860171] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.860374] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 853.860685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 853.861493] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3eaffce1-29db-4391-a891-177706a4277f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.875571] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 853.875571] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f5c4e4-1822-63cf-7c95-fe898f6799e8" [ 853.875571] env[65121]: _type = "Task" [ 853.875571] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.887072] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f5c4e4-1822-63cf-7c95-fe898f6799e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.996841] env[65121]: WARNING openstack [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.997244] env[65121]: WARNING openstack [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.080862] env[65121]: WARNING neutronclient.v2_0.client [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 854.081565] env[65121]: WARNING openstack [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.081919] env[65121]: WARNING openstack [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.258982] env[65121]: DEBUG nova.network.neutron [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Updated VIF entry in instance network info cache for port e904441b-001c-4583-a1e4-98b5fbd6eb74. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 854.258982] env[65121]: DEBUG nova.network.neutron [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Updating instance_info_cache with network_info: [{"id": "e904441b-001c-4583-a1e4-98b5fbd6eb74", "address": "fa:16:3e:5e:ee:13", "network": {"id": "4af4fcd7-26ea-487e-a8e7-90cdec609474", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-431637469-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7d185e5679745caa16ec007c348797b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape904441b-00", "ovs_interfaceid": "e904441b-001c-4583-a1e4-98b5fbd6eb74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 854.336500] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.336500] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.336500] env[65121]: DEBUG nova.network.neutron [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 854.345201] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "refresh_cache-e2e03e5c-4a71-4555-9cd4-d67e178064ec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.345407] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquired lock "refresh_cache-e2e03e5c-4a71-4555-9cd4-d67e178064ec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.345585] env[65121]: DEBUG nova.network.neutron [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 854.393672] env[65121]: INFO nova.compute.manager [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Took 55.91 seconds to build instance. [ 854.403866] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f5c4e4-1822-63cf-7c95-fe898f6799e8, 'name': SearchDatastore_Task, 'duration_secs': 0.016586} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.404206] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.404450] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.404684] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.405071] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.405071] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.405551] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e41b8723-a4bc-49d7-95b8-e0889a7087c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.417437] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.417622] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.418832] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d60e258-e9b8-4400-af55-23afd573ac0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.427115] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 854.427115] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5234533c-ae7d-57a1-5ed8-7d4c957a053f" [ 854.427115] env[65121]: _type = "Task" [ 854.427115] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.443908] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5234533c-ae7d-57a1-5ed8-7d4c957a053f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.529611] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a2b49e-27e4-4d5c-b533-2a3d40b4c3ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.540076] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513397e7-b4b0-4468-9784-1c9716094231 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.577807] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5846f798-f731-4e64-ae37-46e18b39053c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.586951] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a67594-d561-4d22-aba8-3d4150bafc8b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.601884] env[65121]: DEBUG nova.compute.provider_tree [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.765718] env[65121]: DEBUG oslo_concurrency.lockutils [req-13f4ff95-45ff-44ae-a652-89adf36c3fe1 req-52945fc9-2e02-4e9b-84c3-9c9b0a28975a service nova] Releasing lock "refresh_cache-5d352184-7074-4fab-b4ac-cc11a6b936b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.839144] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.839569] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.852324] env[65121]: WARNING neutronclient.v2_0.client [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 854.852965] env[65121]: WARNING openstack [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.853712] env[65121]: WARNING openstack [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.896010] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9c571b2c-2f93-4a05-baa7-b131e854c103 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.544s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.904292] env[65121]: DEBUG nova.network.neutron [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 854.917730] env[65121]: DEBUG nova.network.neutron [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 854.940384] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5234533c-ae7d-57a1-5ed8-7d4c957a053f, 'name': SearchDatastore_Task, 'duration_secs': 0.033064} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.940384] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3dfbc84-a863-4bd7-bc59-dac9c3ad4752 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.948483] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 854.948483] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a7c30f-b841-2e60-4514-671c222625e3" [ 854.948483] env[65121]: _type = "Task" [ 854.948483] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.963209] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a7c30f-b841-2e60-4514-671c222625e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.966189] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.966628] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.983171] env[65121]: DEBUG nova.compute.manager [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Received event network-vif-plugged-027b9031-f921-43b2-8580-de8d028869c3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 854.983171] env[65121]: DEBUG oslo_concurrency.lockutils [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Acquiring lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 854.983171] env[65121]: DEBUG oslo_concurrency.lockutils [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 854.983171] env[65121]: DEBUG oslo_concurrency.lockutils [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.983171] env[65121]: DEBUG nova.compute.manager [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] No waiting events found dispatching network-vif-plugged-027b9031-f921-43b2-8580-de8d028869c3 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 854.983171] env[65121]: WARNING nova.compute.manager [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Received unexpected event network-vif-plugged-027b9031-f921-43b2-8580-de8d028869c3 for instance with vm_state building and task_state spawning. [ 854.983171] env[65121]: DEBUG nova.compute.manager [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Received event network-changed-027b9031-f921-43b2-8580-de8d028869c3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 854.983171] env[65121]: DEBUG nova.compute.manager [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Refreshing instance network info cache due to event network-changed-027b9031-f921-43b2-8580-de8d028869c3. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 854.983171] env[65121]: DEBUG oslo_concurrency.lockutils [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Acquiring lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.022772] env[65121]: DEBUG nova.network.neutron [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 855.063366] env[65121]: WARNING neutronclient.v2_0.client [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.065926] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.065926] env[65121]: WARNING openstack [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.090067] env[65121]: DEBUG nova.compute.manager [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Received event network-changed-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 855.090309] env[65121]: DEBUG nova.compute.manager [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Refreshing instance network info cache due to event network-changed-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 855.090617] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Acquiring lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.090843] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Acquired lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.091090] env[65121]: DEBUG nova.network.neutron [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Refreshing network info cache for port 2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 855.106966] env[65121]: DEBUG nova.scheduler.client.report [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.188190] env[65121]: DEBUG nova.network.neutron [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Updating instance_info_cache with network_info: [{"id": "027b9031-f921-43b2-8580-de8d028869c3", "address": "fa:16:3e:c8:ea:2b", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap027b9031-f9", "ovs_interfaceid": "027b9031-f921-43b2-8580-de8d028869c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 855.399275] env[65121]: DEBUG nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 855.465106] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a7c30f-b841-2e60-4514-671c222625e3, 'name': SearchDatastore_Task, 'duration_secs': 0.049071} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.465337] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.465633] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 5d352184-7074-4fab-b4ac-cc11a6b936b1/5d352184-7074-4fab-b4ac-cc11a6b936b1.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.465948] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-773446e0-b28b-4947-8aa0-e03e18a4228b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.474872] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 855.474872] env[65121]: value = "task-5106575" [ 855.474872] env[65121]: _type = "Task" [ 855.474872] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.489050] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106575, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.529632] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Releasing lock "refresh_cache-e2e03e5c-4a71-4555-9cd4-d67e178064ec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.530443] env[65121]: DEBUG nova.compute.manager [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 855.530786] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.532195] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72141e5a-fbad-4b8d-8fcb-778f81a9ed3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.542244] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.542244] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3219c41e-5bcd-49a0-8657-01ff77922f92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.554051] env[65121]: DEBUG oslo_vmware.api [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 855.554051] env[65121]: value = "task-5106576" [ 855.554051] env[65121]: _type = "Task" [ 855.554051] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.566946] env[65121]: DEBUG oslo_vmware.api [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.598879] env[65121]: WARNING neutronclient.v2_0.client [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.599411] env[65121]: WARNING openstack [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.599725] env[65121]: WARNING openstack [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.614710] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.857s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 855.617733] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.771s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.619847] env[65121]: INFO nova.compute.claims [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.663724] env[65121]: INFO nova.scheduler.client.report [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Deleted allocations for instance cd86f815-4b46-45b1-9a1c-ab2e28f63626 [ 855.691597] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.692674] env[65121]: DEBUG nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Instance network_info: |[{"id": "027b9031-f921-43b2-8580-de8d028869c3", "address": "fa:16:3e:c8:ea:2b", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap027b9031-f9", "ovs_interfaceid": "027b9031-f921-43b2-8580-de8d028869c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 855.694336] env[65121]: DEBUG oslo_concurrency.lockutils [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Acquired lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 855.694645] env[65121]: DEBUG nova.network.neutron [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Refreshing network info cache for port 027b9031-f921-43b2-8580-de8d028869c3 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 855.697971] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:ea:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '027b9031-f921-43b2-8580-de8d028869c3', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.706559] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 855.708396] env[65121]: WARNING neutronclient.v2_0.client [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.709162] env[65121]: WARNING openstack [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.709619] env[65121]: WARNING openstack [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.720814] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.720814] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e6c6d27-11da-4d0d-b62e-a1bf41db1dd8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.750728] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.750728] env[65121]: value = "task-5106577" [ 855.750728] env[65121]: _type = "Task" [ 855.750728] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.763095] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106577, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.842651] env[65121]: WARNING openstack [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.843359] env[65121]: WARNING openstack [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.926513] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.976667] env[65121]: WARNING openstack [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.977149] env[65121]: WARNING openstack [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.003047] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106575, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.010193] env[65121]: WARNING neutronclient.v2_0.client [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.010193] env[65121]: WARNING openstack [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.010193] env[65121]: WARNING openstack [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.074253] env[65121]: DEBUG oslo_vmware.api [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106576, 'name': PowerOffVM_Task, 'duration_secs': 0.183687} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.074637] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.074857] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 856.075243] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd4d90a2-d2b0-4691-bbbc-b72456c1e4dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.104468] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 856.104704] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 856.104889] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Deleting the datastore file [datastore2] e2e03e5c-4a71-4555-9cd4-d67e178064ec {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.105268] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-198c3acd-3740-48a2-b994-caf499ea2593 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.116660] env[65121]: DEBUG oslo_vmware.api [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for the task: (returnval){ [ 856.116660] env[65121]: value = "task-5106579" [ 856.116660] env[65121]: _type = "Task" [ 856.116660] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.136038] env[65121]: DEBUG oslo_vmware.api [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.160054] env[65121]: WARNING neutronclient.v2_0.client [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.160963] env[65121]: WARNING openstack [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.161420] env[65121]: WARNING openstack [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.175233] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a604746f-3185-4e19-bf63-b5256c84c9b2 tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "cd86f815-4b46-45b1-9a1c-ab2e28f63626" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 47.107s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.262870] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106577, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.288822] env[65121]: DEBUG nova.network.neutron [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Updated VIF entry in instance network info cache for port 2e0f93ed-17fe-4ad2-aadd-13b7bdac954c. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 856.290027] env[65121]: DEBUG nova.network.neutron [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Updating instance_info_cache with network_info: [{"id": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "address": "fa:16:3e:80:39:87", "network": {"id": "3e5c7391-4b7a-4d03-ac51-079a45555611", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1489665747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "741e03af2f4d451d9ed77004b644deb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0f93ed-17", "ovs_interfaceid": "2e0f93ed-17fe-4ad2-aadd-13b7bdac954c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.409869] env[65121]: DEBUG nova.network.neutron [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Updated VIF entry in instance network info cache for port 027b9031-f921-43b2-8580-de8d028869c3. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 856.410318] env[65121]: DEBUG nova.network.neutron [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Updating instance_info_cache with network_info: [{"id": "027b9031-f921-43b2-8580-de8d028869c3", "address": "fa:16:3e:c8:ea:2b", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap027b9031-f9", "ovs_interfaceid": "027b9031-f921-43b2-8580-de8d028869c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.503219] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106575, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606014} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.503677] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 5d352184-7074-4fab-b4ac-cc11a6b936b1/5d352184-7074-4fab-b4ac-cc11a6b936b1.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.506047] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.506047] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-641825d3-1e89-4a52-b16a-88003d21820e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.512783] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 856.512783] env[65121]: value = "task-5106580" [ 856.512783] env[65121]: _type = "Task" [ 856.512783] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.532230] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.633325] env[65121]: DEBUG oslo_vmware.api [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Task: {'id': task-5106579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199815} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.634215] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.634215] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.634215] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.634710] env[65121]: INFO nova.compute.manager [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Took 1.10 seconds to destroy the instance on the hypervisor. [ 856.634710] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 856.634877] env[65121]: DEBUG nova.compute.manager [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 856.634877] env[65121]: DEBUG nova.network.neutron [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 856.635148] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.635787] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.636107] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.732033] env[65121]: DEBUG nova.network.neutron [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 856.732033] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.764813] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106577, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.799068] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a030cd1-ab33-4ee8-94cd-f09ba4b6a096 req-ed559cfc-e0ff-4e5a-be87-6b2c475156ea service nova] Releasing lock "refresh_cache-b22ccce7-c54a-4577-9de0-1fd9c10cd189" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.913542] env[65121]: DEBUG oslo_concurrency.lockutils [req-13afcef5-8cb1-43f8-b5c2-96c0bc6d845b req-48b86ef3-af25-4708-90b7-3213d1148e3e service nova] Releasing lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 857.029137] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073727} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.029137] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.029974] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38dc2d8e-40ba-4947-bfce-69cce1d2a14b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.034540] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.034795] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.034997] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.035256] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.035453] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.037858] env[65121]: INFO nova.compute.manager [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Terminating instance [ 857.064764] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 5d352184-7074-4fab-b4ac-cc11a6b936b1/5d352184-7074-4fab-b4ac-cc11a6b936b1.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.068999] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0ce9ab9-4e63-4cad-af9a-838298c97242 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.093768] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 857.093768] env[65121]: value = "task-5106581" [ 857.093768] env[65121]: _type = "Task" [ 857.093768] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.106182] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.237316] env[65121]: DEBUG nova.network.neutron [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 857.267660] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106577, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.338664] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e0a5b2-2a3b-4bc5-8022-61c529984d65 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.348944] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48115255-f37c-446d-9394-1cfdb7231feb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.383386] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0c3ae2-2c7e-4d42-af25-a8e208ff1eec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.395130] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06a5431-fe4c-4341-9cfb-e66de8153007 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.415016] env[65121]: DEBUG nova.compute.provider_tree [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.588880] env[65121]: DEBUG nova.compute.manager [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 857.588880] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 857.589205] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09920b4e-1a8c-479a-9c75-bad64719ccfd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.601870] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.602683] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f382761-468d-47ec-b541-59dbff09f07b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.608355] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106581, 'name': ReconfigVM_Task, 'duration_secs': 0.387937} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.608634] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 5d352184-7074-4fab-b4ac-cc11a6b936b1/5d352184-7074-4fab-b4ac-cc11a6b936b1.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.609337] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-deff5dd8-2caf-49ad-9e8a-004815183a66 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.612319] env[65121]: DEBUG oslo_vmware.api [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 857.612319] env[65121]: value = "task-5106582" [ 857.612319] env[65121]: _type = "Task" [ 857.612319] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.619028] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 857.619028] env[65121]: value = "task-5106583" [ 857.619028] env[65121]: _type = "Task" [ 857.619028] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.623225] env[65121]: DEBUG oslo_vmware.api [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106582, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.635527] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106583, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.742872] env[65121]: INFO nova.compute.manager [-] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Took 1.11 seconds to deallocate network for instance. [ 857.764344] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106577, 'name': CreateVM_Task, 'duration_secs': 1.674541} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.764525] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.765048] env[65121]: WARNING neutronclient.v2_0.client [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 857.765514] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.765700] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 857.766062] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 857.766374] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-548f7315-f2df-47a1-9100-65b8b058bbeb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.772359] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 857.772359] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5261db28-cf1f-0baa-5bb0-dc0ed4921fa5" [ 857.772359] env[65121]: _type = "Task" [ 857.772359] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.782020] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5261db28-cf1f-0baa-5bb0-dc0ed4921fa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.919195] env[65121]: DEBUG nova.scheduler.client.report [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 858.124937] env[65121]: DEBUG oslo_vmware.api [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106582, 'name': PowerOffVM_Task, 'duration_secs': 0.285514} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.128622] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.128832] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 858.129140] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83f106e0-d08a-4b89-b6b0-30bd8ccf9681 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.136840] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106583, 'name': Rename_Task, 'duration_secs': 0.229422} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.138037] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.138037] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39fedf2c-3f97-4968-ae88-0e43a815dbac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.145671] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 858.145671] env[65121]: value = "task-5106585" [ 858.145671] env[65121]: _type = "Task" [ 858.145671] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.155570] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.206994] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.207541] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.207802] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Deleting the datastore file [datastore1] c993d5a4-49d6-43aa-a1f0-4aac91fad953 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.208188] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1845d2fd-a4a4-4071-a738-caa03fff69ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.218238] env[65121]: DEBUG oslo_vmware.api [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for the task: (returnval){ [ 858.218238] env[65121]: value = "task-5106586" [ 858.218238] env[65121]: _type = "Task" [ 858.218238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.230156] env[65121]: DEBUG oslo_vmware.api [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106586, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.250889] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 858.284070] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5261db28-cf1f-0baa-5bb0-dc0ed4921fa5, 'name': SearchDatastore_Task, 'duration_secs': 0.014597} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.285942] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 858.285942] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.285942] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.285942] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.285942] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.285942] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcee5f67-44d6-41e3-8c8a-fb8485edc7f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.297597] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.297597] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.297902] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55b9578f-6f31-456e-b26d-98c9b220101b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.305893] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 858.305893] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ae9a3a-c8de-059e-b753-f1f12f22a1a9" [ 858.305893] env[65121]: _type = "Task" [ 858.305893] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.315809] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ae9a3a-c8de-059e-b753-f1f12f22a1a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.425280] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.808s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 858.425980] env[65121]: DEBUG nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 858.429461] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.846s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 858.431908] env[65121]: INFO nova.compute.claims [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.657214] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106585, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.732440] env[65121]: DEBUG oslo_vmware.api [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Task: {'id': task-5106586, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260599} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.732675] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.732932] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 858.733280] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 858.733524] env[65121]: INFO nova.compute.manager [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Took 1.15 seconds to destroy the instance on the hypervisor. [ 858.733878] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 858.734162] env[65121]: DEBUG nova.compute.manager [-] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 858.734295] env[65121]: DEBUG nova.network.neutron [-] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 858.734759] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.735533] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.735812] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.786280] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.818295] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ae9a3a-c8de-059e-b753-f1f12f22a1a9, 'name': SearchDatastore_Task, 'duration_secs': 0.015572} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.819316] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca6848b0-a78f-4b36-a99a-bfc9a040f63c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.826468] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 858.826468] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5246297f-8da2-3f00-806d-d2c336eab26d" [ 858.826468] env[65121]: _type = "Task" [ 858.826468] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.839512] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246297f-8da2-3f00-806d-d2c336eab26d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.941296] env[65121]: DEBUG nova.compute.utils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 858.942471] env[65121]: DEBUG nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 858.942737] env[65121]: DEBUG nova.network.neutron [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 858.944222] env[65121]: WARNING neutronclient.v2_0.client [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.944638] env[65121]: WARNING neutronclient.v2_0.client [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.945538] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.945717] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.993904] env[65121]: DEBUG nova.policy [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '175d752afbc745258500be8624bfbf14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e491b88dcaad42dcb345b3217bc6f941', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 859.162160] env[65121]: DEBUG oslo_vmware.api [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106585, 'name': PowerOnVM_Task, 'duration_secs': 0.691328} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.164669] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.164669] env[65121]: INFO nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Took 10.11 seconds to spawn the instance on the hypervisor. [ 859.164669] env[65121]: DEBUG nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 859.165751] env[65121]: DEBUG nova.compute.manager [req-94eca9fe-7974-48ef-8c61-937829ace519 req-c4710f1a-c7bc-467a-b5bd-dbc0a23ce210 service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Received event network-vif-deleted-f3966bcc-2651-436f-8bcb-da2f0b915c10 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 859.166212] env[65121]: INFO nova.compute.manager [req-94eca9fe-7974-48ef-8c61-937829ace519 req-c4710f1a-c7bc-467a-b5bd-dbc0a23ce210 service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Neutron deleted interface f3966bcc-2651-436f-8bcb-da2f0b915c10; detaching it from the instance and deleting it from the info cache [ 859.166465] env[65121]: DEBUG nova.network.neutron [req-94eca9fe-7974-48ef-8c61-937829ace519 req-c4710f1a-c7bc-467a-b5bd-dbc0a23ce210 service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 859.168744] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7f87df-4284-482c-aff0-f45d9073a6a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.338514] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246297f-8da2-3f00-806d-d2c336eab26d, 'name': SearchDatastore_Task, 'duration_secs': 0.036265} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.339492] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.339492] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 99b1ff9c-38b7-4947-81bb-d200a3b2a09a/99b1ff9c-38b7-4947-81bb-d200a3b2a09a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.339492] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-131b7f3b-6d33-4296-9499-2132d8f38808 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.349739] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 859.349739] env[65121]: value = "task-5106587" [ 859.349739] env[65121]: _type = "Task" [ 859.349739] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.360519] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.426866] env[65121]: DEBUG nova.network.neutron [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Successfully created port: e7651b37-754f-427b-b5a8-5dc1d76d759c {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 859.453753] env[65121]: DEBUG nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 859.518978] env[65121]: DEBUG nova.network.neutron [-] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 859.676374] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d60f23be-1f44-495c-a0cc-4467aa278dfe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.692921] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be31e75-ed12-441e-8bd6-11182cf11493 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.711939] env[65121]: INFO nova.compute.manager [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Took 53.77 seconds to build instance. [ 859.745666] env[65121]: DEBUG nova.compute.manager [req-94eca9fe-7974-48ef-8c61-937829ace519 req-c4710f1a-c7bc-467a-b5bd-dbc0a23ce210 service nova] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Detach interface failed, port_id=f3966bcc-2651-436f-8bcb-da2f0b915c10, reason: Instance c993d5a4-49d6-43aa-a1f0-4aac91fad953 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 859.861419] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106587, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.025632] env[65121]: INFO nova.compute.manager [-] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Took 1.29 seconds to deallocate network for instance. [ 860.065513] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa81db4-d785-47cf-ad06-6bc1e9cd3a48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.079381] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178f2d49-a82d-49f1-99df-902264f324fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.117338] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e820c16-04e7-42b1-8e94-32d8674ac07c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.129168] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e11750a-5186-43c6-a64a-abd3962bc2a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.147562] env[65121]: DEBUG nova.compute.provider_tree [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.214597] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d65bcf40-382d-48d5-aa4f-cae6ac850de0 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.007s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.363792] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.890115} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.364497] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 99b1ff9c-38b7-4947-81bb-d200a3b2a09a/99b1ff9c-38b7-4947-81bb-d200a3b2a09a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.364641] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.366615] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7dd05ad0-94fa-4537-80c6-dbd317c87626 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.374084] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 860.374084] env[65121]: value = "task-5106588" [ 860.374084] env[65121]: _type = "Task" [ 860.374084] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.386962] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.393514] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5202f174-c044-b118-21f1-13615df5ac06/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 860.394545] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d84892-9fa0-4003-841c-f379085a9624 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.402182] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5202f174-c044-b118-21f1-13615df5ac06/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 860.402368] env[65121]: ERROR oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5202f174-c044-b118-21f1-13615df5ac06/disk-0.vmdk due to incomplete transfer. [ 860.402627] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d1772db3-9029-4c8a-a094-194a76fb6b15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.416696] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5202f174-c044-b118-21f1-13615df5ac06/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 860.416696] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Uploaded image 6b971b17-fd69-4b46-8211-01caca537cc1 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 860.417159] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 860.417380] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-34d8b7b6-acf6-41d0-a436-7a6683e02d57 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.426033] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 860.426033] env[65121]: value = "task-5106589" [ 860.426033] env[65121]: _type = "Task" [ 860.426033] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.434562] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106589, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.467572] env[65121]: DEBUG nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 860.496380] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 860.496636] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 860.496782] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 860.496951] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 860.497099] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 860.497239] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 860.497440] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.497609] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 860.497773] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 860.497941] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 860.498118] env[65121]: DEBUG nova.virt.hardware [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 860.499068] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d456be44-c790-4c16-8320-fa1fa6d49769 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.508574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc9d73c-e3fa-4bab-903c-f74c23608a76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.533859] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.651349] env[65121]: DEBUG nova.scheduler.client.report [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 860.718197] env[65121]: DEBUG nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 860.885702] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088256} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.885980] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.886842] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4079bd9-6965-4e50-97ad-e6a64dd03c2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.920151] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 99b1ff9c-38b7-4947-81bb-d200a3b2a09a/99b1ff9c-38b7-4947-81bb-d200a3b2a09a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.920325] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5552d016-2949-4746-b929-f8a9ad418cf5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.948443] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106589, 'name': Destroy_Task, 'duration_secs': 0.367438} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.950414] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Destroyed the VM [ 860.950681] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 860.951033] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 860.951033] env[65121]: value = "task-5106590" [ 860.951033] env[65121]: _type = "Task" [ 860.951033] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.951246] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ba4c5b63-f6e4-4787-aa12-ea14d548b9ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.962537] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106590, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.964166] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 860.964166] env[65121]: value = "task-5106591" [ 860.964166] env[65121]: _type = "Task" [ 860.964166] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.973615] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106591, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.037537] env[65121]: DEBUG nova.network.neutron [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Successfully updated port: e7651b37-754f-427b-b5a8-5dc1d76d759c {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 861.160305] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.728s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.160305] env[65121]: DEBUG nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 861.162825] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.961s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.164305] env[65121]: DEBUG nova.objects.instance [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lazy-loading 'resources' on Instance uuid 41af3ee5-0ee6-4957-bad0-5aef51d55bd5 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.193337] env[65121]: DEBUG nova.compute.manager [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Received event network-vif-plugged-e7651b37-754f-427b-b5a8-5dc1d76d759c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 861.193721] env[65121]: DEBUG oslo_concurrency.lockutils [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Acquiring lock "93721179-407a-43d9-b0bf-157433a09519-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.193843] env[65121]: DEBUG oslo_concurrency.lockutils [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Lock "93721179-407a-43d9-b0bf-157433a09519-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.194057] env[65121]: DEBUG oslo_concurrency.lockutils [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Lock "93721179-407a-43d9-b0bf-157433a09519-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.194328] env[65121]: DEBUG nova.compute.manager [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] No waiting events found dispatching network-vif-plugged-e7651b37-754f-427b-b5a8-5dc1d76d759c {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 861.194638] env[65121]: WARNING nova.compute.manager [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Received unexpected event network-vif-plugged-e7651b37-754f-427b-b5a8-5dc1d76d759c for instance with vm_state building and task_state spawning. [ 861.195026] env[65121]: DEBUG nova.compute.manager [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Received event network-changed-e7651b37-754f-427b-b5a8-5dc1d76d759c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 861.195274] env[65121]: DEBUG nova.compute.manager [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Refreshing instance network info cache due to event network-changed-e7651b37-754f-427b-b5a8-5dc1d76d759c. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 861.195410] env[65121]: DEBUG oslo_concurrency.lockutils [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Acquiring lock "refresh_cache-93721179-407a-43d9-b0bf-157433a09519" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.195530] env[65121]: DEBUG oslo_concurrency.lockutils [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Acquired lock "refresh_cache-93721179-407a-43d9-b0bf-157433a09519" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.195734] env[65121]: DEBUG nova.network.neutron [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Refreshing network info cache for port e7651b37-754f-427b-b5a8-5dc1d76d759c {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 861.245471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.473806] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106590, 'name': ReconfigVM_Task, 'duration_secs': 0.356638} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.474905] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 99b1ff9c-38b7-4947-81bb-d200a3b2a09a/99b1ff9c-38b7-4947-81bb-d200a3b2a09a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.475165] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4e711a4-d37f-4824-b1ee-52e3a378ccbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.480572] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106591, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.487564] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 861.487564] env[65121]: value = "task-5106592" [ 861.487564] env[65121]: _type = "Task" [ 861.487564] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.499215] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106592, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.540412] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-93721179-407a-43d9-b0bf-157433a09519" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.668493] env[65121]: DEBUG nova.compute.utils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 861.672899] env[65121]: DEBUG nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 861.673228] env[65121]: DEBUG nova.network.neutron [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 861.673522] env[65121]: WARNING neutronclient.v2_0.client [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.673839] env[65121]: WARNING neutronclient.v2_0.client [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.674527] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.675449] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.700440] env[65121]: WARNING neutronclient.v2_0.client [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.700440] env[65121]: WARNING openstack [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.700673] env[65121]: WARNING openstack [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.744621] env[65121]: DEBUG nova.policy [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d7cda02490a471c884b4267218bf234', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2cae196be0ab4b1ab41f653510f7f87a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 861.747253] env[65121]: DEBUG nova.network.neutron [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 861.858853] env[65121]: DEBUG nova.network.neutron [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 861.976936] env[65121]: DEBUG oslo_vmware.api [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106591, 'name': RemoveSnapshot_Task, 'duration_secs': 0.61699} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.979683] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 861.979972] env[65121]: INFO nova.compute.manager [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Took 17.35 seconds to snapshot the instance on the hypervisor. [ 861.998476] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106592, 'name': Rename_Task, 'duration_secs': 0.167759} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.001397] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.001895] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7da455b6-2711-4ba6-adf8-3a7db8dec524 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.010122] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 862.010122] env[65121]: value = "task-5106593" [ 862.010122] env[65121]: _type = "Task" [ 862.010122] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.023574] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.122753] env[65121]: DEBUG nova.network.neutron [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Successfully created port: 296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 862.185908] env[65121]: DEBUG nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 862.298992] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8bb093-944e-4ff8-bdc3-89f5d751b058 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.312668] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8509769-ac24-4c4e-98ef-0e202a835d5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.356965] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2684a18-a5b4-4c80-91c8-1b1d5ed9f8af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.365477] env[65121]: DEBUG oslo_concurrency.lockutils [req-3395ca2b-d033-4db6-a0be-32f39910474f req-250fdb60-1689-4c6a-ad9b-a89f739f8743 service nova] Releasing lock "refresh_cache-93721179-407a-43d9-b0bf-157433a09519" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 862.365947] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-93721179-407a-43d9-b0bf-157433a09519" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.366150] env[65121]: DEBUG nova.network.neutron [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 862.368232] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8877b701-9db2-4a5d-a1a0-6dd6f1733987 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.383901] env[65121]: DEBUG nova.compute.provider_tree [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.523615] env[65121]: DEBUG oslo_vmware.api [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106593, 'name': PowerOnVM_Task, 'duration_secs': 0.508033} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.525934] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.526175] env[65121]: INFO nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Took 10.21 seconds to spawn the instance on the hypervisor. [ 862.526352] env[65121]: DEBUG nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 862.528365] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab9f774-d5b2-4103-9514-667851886e3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.541060] env[65121]: DEBUG nova.compute.manager [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Found 3 images (rotation: 2) {{(pid=65121) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 862.541319] env[65121]: DEBUG nova.compute.manager [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Rotating out 1 backups {{(pid=65121) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5082}} [ 862.541527] env[65121]: DEBUG nova.compute.manager [None req-f276b287-5b2c-4db5-ae9a-e85ee66edbb9 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleting image 1caf99bb-23c5-4c78-a9b4-8ea36bb253c1 {{(pid=65121) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5087}} [ 862.646316] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "5d352184-7074-4fab-b4ac-cc11a6b936b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.646609] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.647061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "5d352184-7074-4fab-b4ac-cc11a6b936b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.647061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.647232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.649654] env[65121]: INFO nova.compute.manager [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Terminating instance [ 862.873636] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.874062] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.887559] env[65121]: DEBUG nova.scheduler.client.report [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 862.951701] env[65121]: DEBUG nova.network.neutron [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 863.005020] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.005020] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.047021] env[65121]: INFO nova.compute.manager [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Took 55.99 seconds to build instance. [ 863.127513] env[65121]: WARNING neutronclient.v2_0.client [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.128350] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.128763] env[65121]: WARNING openstack [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.154111] env[65121]: DEBUG nova.compute.manager [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 863.154111] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.154425] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d769c656-25f1-4fc2-ada4-2bc4ad638dbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.164326] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.164620] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4af73b1d-7a13-4e54-944c-83d13b4b7f1c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.172326] env[65121]: DEBUG oslo_vmware.api [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 863.172326] env[65121]: value = "task-5106594" [ 863.172326] env[65121]: _type = "Task" [ 863.172326] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.181654] env[65121]: DEBUG oslo_vmware.api [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.197484] env[65121]: DEBUG nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 863.230022] env[65121]: DEBUG nova.virt.hardware [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 863.230022] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686effd1-3f83-430b-a60c-c03927ce2f66 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.240355] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75523b89-f967-4ad7-a35c-2072459707c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.299952] env[65121]: DEBUG nova.network.neutron [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Updating instance_info_cache with network_info: [{"id": "e7651b37-754f-427b-b5a8-5dc1d76d759c", "address": "fa:16:3e:01:8c:e8", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7651b37-75", "ovs_interfaceid": "e7651b37-754f-427b-b5a8-5dc1d76d759c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.393536] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.230s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.395969] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.436s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.398117] env[65121]: INFO nova.compute.claims [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.424510] env[65121]: INFO nova.scheduler.client.report [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Deleted allocations for instance 41af3ee5-0ee6-4957-bad0-5aef51d55bd5 [ 863.549249] env[65121]: DEBUG oslo_concurrency.lockutils [None req-52986233-ff56-4f25-be0d-622286756318 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.501s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.684337] env[65121]: DEBUG oslo_vmware.api [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106594, 'name': PowerOffVM_Task, 'duration_secs': 0.250917} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.684595] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.684754] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 863.685017] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-987be38f-f1b4-4e17-b304-83180ed496e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.751422] env[65121]: DEBUG nova.network.neutron [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Successfully updated port: 296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 863.761517] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 863.761748] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 863.762015] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Deleting the datastore file [datastore1] 5d352184-7074-4fab-b4ac-cc11a6b936b1 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.762314] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc4b19c4-cafb-484c-b16c-1ba498d359f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.774807] env[65121]: DEBUG oslo_vmware.api [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for the task: (returnval){ [ 863.774807] env[65121]: value = "task-5106596" [ 863.774807] env[65121]: _type = "Task" [ 863.774807] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.786284] env[65121]: DEBUG oslo_vmware.api [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.800908] env[65121]: DEBUG nova.compute.manager [req-545af9e7-9c94-4a44-9844-e25e8a2defb9 req-e9107b3b-3722-403e-963a-d9ef7f780180 service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Received event network-vif-plugged-296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 863.801173] env[65121]: DEBUG oslo_concurrency.lockutils [req-545af9e7-9c94-4a44-9844-e25e8a2defb9 req-e9107b3b-3722-403e-963a-d9ef7f780180 service nova] Acquiring lock "7fcab9cb-9a0f-4a50-9244-687889443a31-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.801550] env[65121]: DEBUG oslo_concurrency.lockutils [req-545af9e7-9c94-4a44-9844-e25e8a2defb9 req-e9107b3b-3722-403e-963a-d9ef7f780180 service nova] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.801725] env[65121]: DEBUG oslo_concurrency.lockutils [req-545af9e7-9c94-4a44-9844-e25e8a2defb9 req-e9107b3b-3722-403e-963a-d9ef7f780180 service nova] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.801887] env[65121]: DEBUG nova.compute.manager [req-545af9e7-9c94-4a44-9844-e25e8a2defb9 req-e9107b3b-3722-403e-963a-d9ef7f780180 service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] No waiting events found dispatching network-vif-plugged-296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 863.802054] env[65121]: WARNING nova.compute.manager [req-545af9e7-9c94-4a44-9844-e25e8a2defb9 req-e9107b3b-3722-403e-963a-d9ef7f780180 service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Received unexpected event network-vif-plugged-296c4d11-0b03-45a7-8019-6f23e26fea6e for instance with vm_state building and task_state spawning. [ 863.802851] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-93721179-407a-43d9-b0bf-157433a09519" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.803318] env[65121]: DEBUG nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance network_info: |[{"id": "e7651b37-754f-427b-b5a8-5dc1d76d759c", "address": "fa:16:3e:01:8c:e8", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7651b37-75", "ovs_interfaceid": "e7651b37-754f-427b-b5a8-5dc1d76d759c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 863.803764] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:8c:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7651b37-754f-427b-b5a8-5dc1d76d759c', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.812138] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating folder: Project (e491b88dcaad42dcb345b3217bc6f941). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.813172] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fa95af7-1772-4473-9072-46b946d3cb23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.827753] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created folder: Project (e491b88dcaad42dcb345b3217bc6f941) in parent group-v993268. [ 863.828117] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating folder: Instances. Parent ref: group-v993446. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.828500] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-336f1d43-220b-43fb-92bc-0a64ad036c73 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.841346] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created folder: Instances in parent group-v993446. [ 863.841633] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 863.841836] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 863.842079] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4061215f-aa22-4fc6-812c-5899da957dcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.863712] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.863712] env[65121]: value = "task-5106599" [ 863.863712] env[65121]: _type = "Task" [ 863.863712] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.872473] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106599, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.934429] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c2c0dfd3-0ce8-4330-89c1-4b48148bf347 tempest-ServerTagsTestJSON-1810896626 tempest-ServerTagsTestJSON-1810896626-project-member] Lock "41af3ee5-0ee6-4957-bad0-5aef51d55bd5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.111s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.051611] env[65121]: DEBUG nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 864.123681] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Acquiring lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.123681] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Acquired lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.123681] env[65121]: DEBUG nova.network.neutron [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 864.255037] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.255037] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquired lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.255521] env[65121]: DEBUG nova.network.neutron [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 864.285760] env[65121]: DEBUG oslo_vmware.api [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Task: {'id': task-5106596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217814} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.286028] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.286237] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.286478] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.286672] env[65121]: INFO nova.compute.manager [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Took 1.13 seconds to destroy the instance on the hypervisor. [ 864.286912] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 864.287118] env[65121]: DEBUG nova.compute.manager [-] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 864.287286] env[65121]: DEBUG nova.network.neutron [-] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 864.287638] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.288258] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.289205] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.377328] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106599, 'name': CreateVM_Task, 'duration_secs': 0.414465} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.379090] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.380747] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.384069] env[65121]: WARNING neutronclient.v2_0.client [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.384069] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.384069] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.384069] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 864.384069] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcfd44c6-39fd-4029-9291-9829496c14ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.390439] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 864.390439] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5248ca3c-abfe-3bd3-9617-d24f4bc7a1e6" [ 864.390439] env[65121]: _type = "Task" [ 864.390439] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.404569] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5248ca3c-abfe-3bd3-9617-d24f4bc7a1e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.578493] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.627135] env[65121]: WARNING neutronclient.v2_0.client [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.627848] env[65121]: WARNING openstack [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.629449] env[65121]: WARNING openstack [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.757119] env[65121]: WARNING openstack [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.757729] env[65121]: WARNING openstack [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.768766] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.769281] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.894386] env[65121]: DEBUG nova.network.neutron [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 864.907671] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5248ca3c-abfe-3bd3-9617-d24f4bc7a1e6, 'name': SearchDatastore_Task, 'duration_secs': 0.011603} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.910350] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.910586] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.910807] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.910945] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.911131] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.911570] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8be287a1-9ad1-4442-ac01-9487a86013de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.929101] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.929281] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.930054] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a94aece2-dbf7-4c61-8ffc-3200fd14a1f3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.938300] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 864.938300] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5230c40f-40a1-7348-5c90-29726e53f115" [ 864.938300] env[65121]: _type = "Task" [ 864.938300] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.946310] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5230c40f-40a1-7348-5c90-29726e53f115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.982761] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.983133] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.991359] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea73d96f-4dc9-4ce9-a468-02729ab3b79e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.999340] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7d9a26-e886-4c92-97be-2ef9d38c0f84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.031771] env[65121]: WARNING neutronclient.v2_0.client [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.032519] env[65121]: WARNING openstack [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.032785] env[65121]: WARNING openstack [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.041328] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0a9498-7bcb-4eb4-bafa-ebad4321f93c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.054499] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce43542-5c23-4e5b-95df-889ef0849047 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.070771] env[65121]: DEBUG nova.compute.provider_tree [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.124089] env[65121]: WARNING neutronclient.v2_0.client [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.124974] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.125193] env[65121]: WARNING openstack [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.143607] env[65121]: DEBUG nova.network.neutron [-] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.184380] env[65121]: DEBUG nova.network.neutron [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Updating instance_info_cache with network_info: [{"id": "027b9031-f921-43b2-8580-de8d028869c3", "address": "fa:16:3e:c8:ea:2b", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap027b9031-f9", "ovs_interfaceid": "027b9031-f921-43b2-8580-de8d028869c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.291991] env[65121]: DEBUG nova.network.neutron [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Updating instance_info_cache with network_info: [{"id": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "address": "fa:16:3e:b7:7c:fa", "network": {"id": "0004f91f-e9bd-4c04-b81e-2709bf0dbceb", "bridge": "br-int", "label": "tempest-ServersTestJSON-1838786777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cae196be0ab4b1ab41f653510f7f87a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap296c4d11-0b", "ovs_interfaceid": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.338867] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.339137] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.339320] env[65121]: DEBUG nova.compute.manager [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 865.340262] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9f619c-a69c-4fce-b770-82bb57a985e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.347503] env[65121]: DEBUG nova.compute.manager [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 865.348243] env[65121]: DEBUG nova.objects.instance [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'flavor' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.452825] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5230c40f-40a1-7348-5c90-29726e53f115, 'name': SearchDatastore_Task, 'duration_secs': 0.026733} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.454012] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-055eef6d-2772-4a5e-a50b-21b207241541 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.460745] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 865.460745] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52710dff-8382-cd75-d1ea-262c3ff80f53" [ 865.460745] env[65121]: _type = "Task" [ 865.460745] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.473368] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52710dff-8382-cd75-d1ea-262c3ff80f53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.577202] env[65121]: DEBUG nova.scheduler.client.report [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 865.646340] env[65121]: INFO nova.compute.manager [-] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Took 1.36 seconds to deallocate network for instance. [ 865.687065] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Releasing lock "refresh_cache-99b1ff9c-38b7-4947-81bb-d200a3b2a09a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.687338] env[65121]: DEBUG nova.compute.manager [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Inject network info {{(pid=65121) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 865.687663] env[65121]: DEBUG nova.compute.manager [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] network_info to inject: |[{"id": "027b9031-f921-43b2-8580-de8d028869c3", "address": "fa:16:3e:c8:ea:2b", "network": {"id": "54e5ee3a-6452-4caf-97c4-5525d871fce1", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-321699549-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "083c4f2aa7bf4167b472db6bebb9d8bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap027b9031-f9", "ovs_interfaceid": "027b9031-f921-43b2-8580-de8d028869c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 865.693677] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Reconfiguring VM instance to set the machine id {{(pid=65121) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 865.695936] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9df9d085-6fe2-450c-9f48-68e23ba7d742 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.716609] env[65121]: DEBUG oslo_vmware.api [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Waiting for the task: (returnval){ [ 865.716609] env[65121]: value = "task-5106600" [ 865.716609] env[65121]: _type = "Task" [ 865.716609] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.727986] env[65121]: DEBUG oslo_vmware.api [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Task: {'id': task-5106600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.797032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Releasing lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.797032] env[65121]: DEBUG nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Instance network_info: |[{"id": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "address": "fa:16:3e:b7:7c:fa", "network": {"id": "0004f91f-e9bd-4c04-b81e-2709bf0dbceb", "bridge": "br-int", "label": "tempest-ServersTestJSON-1838786777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cae196be0ab4b1ab41f653510f7f87a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap296c4d11-0b", "ovs_interfaceid": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 865.797032] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:7c:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '777870ab-362f-4a17-9c1c-8d9cc26cd4ce', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '296c4d11-0b03-45a7-8019-6f23e26fea6e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.804064] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Creating folder: Project (2cae196be0ab4b1ab41f653510f7f87a). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.804384] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-243b5405-a171-4aa2-a667-aba7fad21fdf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.819162] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Created folder: Project (2cae196be0ab4b1ab41f653510f7f87a) in parent group-v993268. [ 865.819162] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Creating folder: Instances. Parent ref: group-v993449. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.819162] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65cdd19d-3499-4bf3-a8c0-c3d4e3acd1aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.832748] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Created folder: Instances in parent group-v993449. [ 865.832748] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 865.832748] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.832748] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ac2ded0-7a7b-4afe-875a-f66edca7928b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.857174] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.857174] env[65121]: value = "task-5106603" [ 865.857174] env[65121]: _type = "Task" [ 865.857174] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.869851] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106603, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.973755] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52710dff-8382-cd75-d1ea-262c3ff80f53, 'name': SearchDatastore_Task, 'duration_secs': 0.030724} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.974594] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.974594] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 865.976129] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-909049ed-eb4e-4004-b909-c98e627f8375 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.983899] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 865.983899] env[65121]: value = "task-5106604" [ 865.983899] env[65121]: _type = "Task" [ 865.983899] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.999552] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.002869] env[65121]: DEBUG nova.compute.manager [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Received event network-changed-296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 866.003265] env[65121]: DEBUG nova.compute.manager [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Refreshing instance network info cache due to event network-changed-296c4d11-0b03-45a7-8019-6f23e26fea6e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 866.003429] env[65121]: DEBUG oslo_concurrency.lockutils [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Acquiring lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.003609] env[65121]: DEBUG oslo_concurrency.lockutils [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Acquired lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.003848] env[65121]: DEBUG nova.network.neutron [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Refreshing network info cache for port 296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 866.084024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 866.084024] env[65121]: DEBUG nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 866.085796] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.011s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.086208] env[65121]: DEBUG nova.objects.instance [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lazy-loading 'resources' on Instance uuid f694fcb6-053b-4649-ac63-7fa98b1373eb {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.154533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.230084] env[65121]: DEBUG oslo_vmware.api [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] Task: {'id': task-5106600, 'name': ReconfigVM_Task, 'duration_secs': 0.326489} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.230427] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba0b136-00d5-4d16-aa9d-f753cc75b857 tempest-ServersAdminTestJSON-1502023876 tempest-ServersAdminTestJSON-1502023876-project-admin] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Reconfigured VM instance to set the machine id {{(pid=65121) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 866.361023] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.361023] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-adfed420-5fe6-439f-b3f2-969b836164ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.374311] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106603, 'name': CreateVM_Task, 'duration_secs': 0.49302} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.375931] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.376752] env[65121]: DEBUG oslo_vmware.api [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 866.376752] env[65121]: value = "task-5106605" [ 866.376752] env[65121]: _type = "Task" [ 866.376752] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.377414] env[65121]: WARNING neutronclient.v2_0.client [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.377723] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.378084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.378804] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 866.378869] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9f257fc-fc36-4a50-a078-cdb72702223a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.389719] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 866.389719] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5248cf77-272f-6fef-35f2-d47072382437" [ 866.389719] env[65121]: _type = "Task" [ 866.389719] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.393638] env[65121]: DEBUG oslo_vmware.api [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106605, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.404583] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5248cf77-272f-6fef-35f2-d47072382437, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.499210] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106604, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.507146] env[65121]: WARNING neutronclient.v2_0.client [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.508030] env[65121]: WARNING openstack [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.509736] env[65121]: WARNING openstack [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.591460] env[65121]: DEBUG nova.compute.utils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 866.593123] env[65121]: DEBUG nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 866.593323] env[65121]: DEBUG nova.network.neutron [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 866.593745] env[65121]: WARNING neutronclient.v2_0.client [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.596039] env[65121]: WARNING neutronclient.v2_0.client [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.596039] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.596039] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.656338] env[65121]: DEBUG nova.policy [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a602d2a28164f069851eb46c57f47ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '693c1c83a1d5453bbedaf8ce2bf7c8d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 866.673862] env[65121]: WARNING openstack [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.673862] env[65121]: WARNING openstack [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.749599] env[65121]: WARNING neutronclient.v2_0.client [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.750075] env[65121]: WARNING openstack [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.750507] env[65121]: WARNING openstack [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.889870] env[65121]: DEBUG nova.network.neutron [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Updated VIF entry in instance network info cache for port 296c4d11-0b03-45a7-8019-6f23e26fea6e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 866.890290] env[65121]: DEBUG nova.network.neutron [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Updating instance_info_cache with network_info: [{"id": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "address": "fa:16:3e:b7:7c:fa", "network": {"id": "0004f91f-e9bd-4c04-b81e-2709bf0dbceb", "bridge": "br-int", "label": "tempest-ServersTestJSON-1838786777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cae196be0ab4b1ab41f653510f7f87a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap296c4d11-0b", "ovs_interfaceid": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.902999] env[65121]: DEBUG oslo_vmware.api [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106605, 'name': PowerOffVM_Task, 'duration_secs': 0.476847} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.904188] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.904390] env[65121]: DEBUG nova.compute.manager [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 866.905189] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23feced6-6a5e-4d2c-b7c5-c0331dad547e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.916172] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5248cf77-272f-6fef-35f2-d47072382437, 'name': SearchDatastore_Task, 'duration_secs': 0.058534} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.917300] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.917425] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.917962] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.917962] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.917962] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.918505] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35b9a7ea-de20-4efa-af0d-6b536056dfbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.940569] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.940765] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.941889] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffe164d8-e969-4573-afe2-619ccf4390be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.952217] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 866.952217] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5294e18b-ba6f-a1d6-3158-ecfc08d72ffb" [ 866.952217] env[65121]: _type = "Task" [ 866.952217] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.961288] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5294e18b-ba6f-a1d6-3158-ecfc08d72ffb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.996216] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106604, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587345} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.996565] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.996803] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.997073] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-517c5273-5ad7-4ac0-87a7-d1e741228e4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.004069] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 867.004069] env[65121]: value = "task-5106606" [ 867.004069] env[65121]: _type = "Task" [ 867.004069] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.015571] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106606, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.041491] env[65121]: DEBUG nova.network.neutron [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Successfully created port: 33fba103-6efa-42f4-9689-339edc299410 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 867.107179] env[65121]: DEBUG nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 867.234819] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604da27d-60ad-449b-9fdc-6bb4a17a1d8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.243896] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f14f2a-0e57-469a-bb79-3331e0e1f425 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.276846] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0682614-26f2-4090-b4cc-15f56a21a409 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.286585] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b195b2-d080-43f9-8bb6-c435b84f9663 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.305362] env[65121]: DEBUG nova.compute.provider_tree [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.395968] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-cc002961-d742-4255-88c7-f0a5a39424b1-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.396320] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-cc002961-d742-4255-88c7-f0a5a39424b1-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.396479] env[65121]: DEBUG nova.objects.instance [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'flavor' on Instance uuid cc002961-d742-4255-88c7-f0a5a39424b1 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.398073] env[65121]: DEBUG oslo_concurrency.lockutils [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] Releasing lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.398409] env[65121]: DEBUG nova.compute.manager [req-57826aa9-cae3-4249-a643-eea44eb72422 req-750d368b-dc02-4c5a-bfcc-4f1a51ee70ca service nova] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Received event network-vif-deleted-e904441b-001c-4583-a1e4-98b5fbd6eb74 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 867.432153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e1482f5c-575d-4583-9923-22eaa5c99deb tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.093s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.463535] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5294e18b-ba6f-a1d6-3158-ecfc08d72ffb, 'name': SearchDatastore_Task, 'duration_secs': 0.026133} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.464377] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-337bab7b-814c-4ccc-b4df-7da20e9fe1a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.471744] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 867.471744] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f31bb1-c128-b81d-92ec-7a27c038eaaa" [ 867.471744] env[65121]: _type = "Task" [ 867.471744] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.479570] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f31bb1-c128-b81d-92ec-7a27c038eaaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.514572] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106606, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077559} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.514868] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.515707] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b446f2b3-dec4-4c27-aa80-a6c18c9c1427 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.540976] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.541383] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0dd595e-b442-4553-905d-4eccf27f8863 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.563277] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 867.563277] env[65121]: value = "task-5106607" [ 867.563277] env[65121]: _type = "Task" [ 867.563277] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.571952] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106607, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.809181] env[65121]: DEBUG nova.scheduler.client.report [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 867.901505] env[65121]: WARNING neutronclient.v2_0.client [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.902249] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.902604] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.983952] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f31bb1-c128-b81d-92ec-7a27c038eaaa, 'name': SearchDatastore_Task, 'duration_secs': 0.057054} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.984278] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.984542] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 7fcab9cb-9a0f-4a50-9244-687889443a31/7fcab9cb-9a0f-4a50-9244-687889443a31.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.984828] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-342adaee-8008-49fe-973c-3b0271fb4205 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.993607] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 867.993607] env[65121]: value = "task-5106608" [ 867.993607] env[65121]: _type = "Task" [ 867.993607] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.003996] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106608, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.019123] env[65121]: DEBUG nova.objects.instance [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'pci_requests' on Instance uuid cc002961-d742-4255-88c7-f0a5a39424b1 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.076648] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106607, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.119450] env[65121]: DEBUG nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 868.157373] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 868.157637] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 868.157804] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 868.157972] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 868.158136] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 868.158387] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 868.158603] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.158777] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 868.158976] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 868.159143] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 868.159306] env[65121]: DEBUG nova.virt.hardware [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 868.160249] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311c626d-cd00-4f85-957b-4b153e2359b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.172974] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7af17c4-d432-468a-87d7-ed79384225d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.314842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.229s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.317680] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.753s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.319411] env[65121]: INFO nova.compute.claims [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.347745] env[65121]: INFO nova.scheduler.client.report [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Deleted allocations for instance f694fcb6-053b-4649-ac63-7fa98b1373eb [ 868.506990] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106608, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.522748] env[65121]: DEBUG nova.objects.base [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 868.523197] env[65121]: DEBUG nova.network.neutron [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 868.526962] env[65121]: WARNING neutronclient.v2_0.client [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.526962] env[65121]: WARNING neutronclient.v2_0.client [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.526962] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.526962] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.575606] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106607, 'name': ReconfigVM_Task, 'duration_secs': 0.899073} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.575978] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.576698] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57037ab9-aa56-4e24-8156-6ed7cc615102 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.586104] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 868.586104] env[65121]: value = "task-5106609" [ 868.586104] env[65121]: _type = "Task" [ 868.586104] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.593647] env[65121]: DEBUG nova.policy [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 868.601055] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106609, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.676072] env[65121]: DEBUG nova.network.neutron [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Successfully updated port: 33fba103-6efa-42f4-9689-339edc299410 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 868.855609] env[65121]: DEBUG oslo_concurrency.lockutils [None req-45056dd1-150d-4a22-8374-c4146ee31697 tempest-ServerExternalEventsTest-1013131577 tempest-ServerExternalEventsTest-1013131577-project-member] Lock "f694fcb6-053b-4649-ac63-7fa98b1373eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.803s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.936040] env[65121]: DEBUG nova.network.neutron [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Successfully created port: c20059cc-f3b0-4686-acf7-f170edf7da11 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 869.005158] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106608, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.097108] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106609, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.179580] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.179758] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.179931] env[65121]: DEBUG nova.network.neutron [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 869.505527] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106608, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.580141] env[65121]: DEBUG nova.compute.manager [req-dcc18256-96bc-4d94-90b2-60238969b270 req-c9a098d1-dfc1-4f4a-bc79-8542d8e71480 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Received event network-vif-plugged-33fba103-6efa-42f4-9689-339edc299410 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 869.580141] env[65121]: DEBUG oslo_concurrency.lockutils [req-dcc18256-96bc-4d94-90b2-60238969b270 req-c9a098d1-dfc1-4f4a-bc79-8542d8e71480 service nova] Acquiring lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.580517] env[65121]: DEBUG oslo_concurrency.lockutils [req-dcc18256-96bc-4d94-90b2-60238969b270 req-c9a098d1-dfc1-4f4a-bc79-8542d8e71480 service nova] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.580517] env[65121]: DEBUG oslo_concurrency.lockutils [req-dcc18256-96bc-4d94-90b2-60238969b270 req-c9a098d1-dfc1-4f4a-bc79-8542d8e71480 service nova] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.580517] env[65121]: DEBUG nova.compute.manager [req-dcc18256-96bc-4d94-90b2-60238969b270 req-c9a098d1-dfc1-4f4a-bc79-8542d8e71480 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] No waiting events found dispatching network-vif-plugged-33fba103-6efa-42f4-9689-339edc299410 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 869.580726] env[65121]: WARNING nova.compute.manager [req-dcc18256-96bc-4d94-90b2-60238969b270 req-c9a098d1-dfc1-4f4a-bc79-8542d8e71480 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Received unexpected event network-vif-plugged-33fba103-6efa-42f4-9689-339edc299410 for instance with vm_state building and task_state spawning. [ 869.601719] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106609, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.683201] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.683879] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.725532] env[65121]: DEBUG nova.network.neutron [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 869.759247] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.759627] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.808269] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3889fe49-4f38-46cd-b8d3-f3b7be499ca9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.816731] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb85aca-1828-4ee6-897c-aade95a4a38d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.852535] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23e2215-671d-49f8-90d4-32e4bd1d8907 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.862435] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875fd622-133c-43d9-9773-3dde95a59af7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.877600] env[65121]: DEBUG nova.compute.provider_tree [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.880069] env[65121]: WARNING neutronclient.v2_0.client [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.880820] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.881228] env[65121]: WARNING openstack [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.890715] env[65121]: DEBUG nova.scheduler.client.report [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.982055] env[65121]: DEBUG nova.network.neutron [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Updating instance_info_cache with network_info: [{"id": "33fba103-6efa-42f4-9689-339edc299410", "address": "fa:16:3e:4a:d6:b4", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33fba103-6e", "ovs_interfaceid": "33fba103-6efa-42f4-9689-339edc299410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 870.007567] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106608, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.598296} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.007843] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 7fcab9cb-9a0f-4a50-9244-687889443a31/7fcab9cb-9a0f-4a50-9244-687889443a31.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.008066] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.008321] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b6f5a07-f52c-4319-a9b7-60e232b8d391 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.014670] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 870.014670] env[65121]: value = "task-5106610" [ 870.014670] env[65121]: _type = "Task" [ 870.014670] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.027533] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.098053] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106609, 'name': Rename_Task, 'duration_secs': 1.440034} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.098053] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.098053] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a57ac583-77de-490b-a498-c903347fd725 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.104610] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 870.104610] env[65121]: value = "task-5106611" [ 870.104610] env[65121]: _type = "Task" [ 870.104610] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.116243] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.397040] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.397228] env[65121]: DEBUG nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 870.400968] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.719s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 870.401428] env[65121]: DEBUG nova.objects.instance [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lazy-loading 'resources' on Instance uuid b1ad9d15-7ef4-404a-9751-ecdd02ea768b {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.485319] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 870.485716] env[65121]: DEBUG nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Instance network_info: |[{"id": "33fba103-6efa-42f4-9689-339edc299410", "address": "fa:16:3e:4a:d6:b4", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33fba103-6e", "ovs_interfaceid": "33fba103-6efa-42f4-9689-339edc299410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 870.486367] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:d6:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33fba103-6efa-42f4-9689-339edc299410', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.495254] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 870.495827] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.496106] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eec55e7b-c770-48db-93d3-8529d817ec7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.520343] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.520343] env[65121]: value = "task-5106612" [ 870.520343] env[65121]: _type = "Task" [ 870.520343] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.528164] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072497} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.528164] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.528961] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c04e229-487a-4f9b-af44-31146d78120b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.535069] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106612, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.559435] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 7fcab9cb-9a0f-4a50-9244-687889443a31/7fcab9cb-9a0f-4a50-9244-687889443a31.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.560770] env[65121]: DEBUG nova.network.neutron [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Successfully updated port: c20059cc-f3b0-4686-acf7-f170edf7da11 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 870.562434] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9cf94bf-cad7-4c2d-b2cd-6456e2cdd72a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.579595] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.579842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 870.579963] env[65121]: DEBUG nova.network.neutron [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 870.588952] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 870.588952] env[65121]: value = "task-5106613" [ 870.588952] env[65121]: _type = "Task" [ 870.588952] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.600866] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.615237] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106611, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.750164] env[65121]: DEBUG nova.compute.manager [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Stashing vm_state: stopped {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 870.905523] env[65121]: DEBUG nova.compute.utils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 870.912340] env[65121]: DEBUG nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 870.912340] env[65121]: DEBUG nova.network.neutron [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 870.912340] env[65121]: WARNING neutronclient.v2_0.client [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.912340] env[65121]: WARNING neutronclient.v2_0.client [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.913557] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.913557] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.972661] env[65121]: DEBUG nova.policy [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f8ab21ef27c452d99654587cb0e1cc7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c6912ada4ba462fb669f0ac353e96ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 871.032742] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106612, 'name': CreateVM_Task, 'duration_secs': 0.350198} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.035491] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.037140] env[65121]: WARNING neutronclient.v2_0.client [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.038474] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.038474] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.038474] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 871.038474] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-622cefa4-1627-4776-8975-1f4538fb43e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.044847] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 871.044847] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525e7b8a-aa60-f843-a9c1-a6e3bd359e63" [ 871.044847] env[65121]: _type = "Task" [ 871.044847] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.060305] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525e7b8a-aa60-f843-a9c1-a6e3bd359e63, 'name': SearchDatastore_Task, 'duration_secs': 0.010687} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.060634] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 871.060882] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.061151] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.061295] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.061490] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.061776] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97b7eebe-1930-4918-9b5b-994a27ebb673 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.072514] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.072694] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.073830] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e64b8261-cbe6-4ff5-94b6-3345fecca8ea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.080466] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 871.080466] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526e1b5f-a7d0-a1c2-9a7e-1e0f74394209" [ 871.080466] env[65121]: _type = "Task" [ 871.080466] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.088459] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.088849] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.107019] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526e1b5f-a7d0-a1c2-9a7e-1e0f74394209, 'name': SearchDatastore_Task, 'duration_secs': 0.012501} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.107019] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0a59ed4-f276-4a32-8e77-3fef269141b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.116181] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106613, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.126823] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106611, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.127184] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 871.127184] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524f73e9-7fa1-cf38-1ab5-4acf5bbb7e6a" [ 871.127184] env[65121]: _type = "Task" [ 871.127184] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.140105] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524f73e9-7fa1-cf38-1ab5-4acf5bbb7e6a, 'name': SearchDatastore_Task, 'duration_secs': 0.01047} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.140322] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 871.140603] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/8a23da27-ce6c-4453-9036-65eeeda3ce0d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.141205] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12048b0a-27bf-4783-97bb-13a3c1e79e10 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.152436] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 871.152436] env[65121]: value = "task-5106614" [ 871.152436] env[65121]: _type = "Task" [ 871.152436] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.164483] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106614, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.281623] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.285721] env[65121]: WARNING nova.network.neutron [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] 5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd already exists in list: networks containing: ['5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd']. ignoring it [ 871.370148] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.371068] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.411909] env[65121]: DEBUG nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 871.432566] env[65121]: DEBUG nova.network.neutron [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Successfully created port: 88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 871.528491] env[65121]: WARNING neutronclient.v2_0.client [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.528789] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.529186] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.610033] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106613, 'name': ReconfigVM_Task, 'duration_secs': 0.750206} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.615076] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 7fcab9cb-9a0f-4a50-9244-687889443a31/7fcab9cb-9a0f-4a50-9244-687889443a31.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.615076] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52db9456-16cf-4661-9e50-e93fef1935f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.620460] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4b167c-6f06-43c0-9ed1-eb6f083937fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.629577] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106611, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.631797] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 871.631797] env[65121]: value = "task-5106615" [ 871.631797] env[65121]: _type = "Task" [ 871.631797] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.639010] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8dbbcd-2037-4c0c-ae80-fc7a66b3b5db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.646506] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106615, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.682627] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa0ae6c-feb0-4cdb-abee-4e574b9381cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.693411] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106614, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.703770] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e92c3e-b033-4190-875c-44efef96b24f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.723475] env[65121]: DEBUG nova.compute.provider_tree [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.859725] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.860126] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.920933] env[65121]: INFO nova.virt.block_device [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Booting with volume acaa084b-972e-4290-a0b1-dfd2ae6ff4da at /dev/sda [ 871.975863] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0923f04f-a750-4bb2-91a8-a8806431a8c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.993079] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09841146-145d-4f7e-b401-e3d61f7c378e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.005414] env[65121]: WARNING neutronclient.v2_0.client [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 872.006093] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 872.006465] env[65121]: WARNING openstack [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.044291] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-044a4408-e340-4d2f-ac69-a6f9e5e16028 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.058687] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c42c3aa-435c-43bc-a218-4dd1ba08c5a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.101435] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fd056d-3cbe-4aca-adb5-38db2c63c62d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.109629] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9a957b-d099-4fa5-b0e8-4cecf54e73cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.121912] env[65121]: DEBUG nova.network.neutron [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c20059cc-f3b0-4686-acf7-f170edf7da11", "address": "fa:16:3e:07:36:ec", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc20059cc-f3", "ovs_interfaceid": "c20059cc-f3b0-4686-acf7-f170edf7da11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 872.130801] env[65121]: DEBUG oslo_vmware.api [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106611, 'name': PowerOnVM_Task, 'duration_secs': 1.589959} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.135097] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.135458] env[65121]: INFO nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Took 11.67 seconds to spawn the instance on the hypervisor. [ 872.135661] env[65121]: DEBUG nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 872.136157] env[65121]: DEBUG nova.virt.block_device [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updating existing volume attachment record: 1e1db2d6-cbba-4b2e-92e7-e72e30b95fd4 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 872.141169] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e7e832-0bef-48e1-9854-441a4cbf3980 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.161055] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106615, 'name': Rename_Task, 'duration_secs': 0.23107} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.163330] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.166801] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5f35666-762b-4d64-9d0f-b48ad4048f07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.176290] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 872.176290] env[65121]: value = "task-5106616" [ 872.176290] env[65121]: _type = "Task" [ 872.176290] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.199437] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106616, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.206345] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106614, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566641} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.206757] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/8a23da27-ce6c-4453-9036-65eeeda3ce0d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.207085] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.207474] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d5683236-9b46-4a59-9f35-af3d0ec87db4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.221251] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 872.221251] env[65121]: value = "task-5106617" [ 872.221251] env[65121]: _type = "Task" [ 872.221251] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.228531] env[65121]: DEBUG nova.scheduler.client.report [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 872.241362] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106617, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.625604] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.626343] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.627134] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 872.627366] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f79f7a-bde8-43cd-9df8-24d19c49d49e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.645298] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 872.646026] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 872.646026] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 872.646026] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 872.646026] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 872.646240] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 872.646348] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.646497] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 872.646658] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 872.646946] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 872.647028] env[65121]: DEBUG nova.virt.hardware [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 872.654026] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Reconfiguring VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 872.654239] env[65121]: INFO nova.compute.manager [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Rebuilding instance [ 872.656427] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4560c6e9-5bd5-4a1f-a25c-173397dfad45 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.677648] env[65121]: DEBUG oslo_vmware.api [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 872.677648] env[65121]: value = "task-5106618" [ 872.677648] env[65121]: _type = "Task" [ 872.677648] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.678550] env[65121]: INFO nova.compute.manager [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Took 57.86 seconds to build instance. [ 872.698826] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106616, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.698826] env[65121]: DEBUG oslo_vmware.api [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106618, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.732967] env[65121]: DEBUG nova.compute.manager [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 872.733893] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bc908c-992f-4386-99bb-d053c68394c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.740280] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.339s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.742674] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106617, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.138974} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.743117] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.736s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 872.744623] env[65121]: INFO nova.compute.claims [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.747767] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.749163] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db2be0f-6aa5-4cb0-abab-94472b30ac23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.777838] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/8a23da27-ce6c-4453-9036-65eeeda3ce0d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.779104] env[65121]: INFO nova.scheduler.client.report [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleted allocations for instance b1ad9d15-7ef4-404a-9751-ecdd02ea768b [ 872.780458] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c89c8be1-95b4-46e8-a7e4-4484e0c2566d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.809867] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 872.809867] env[65121]: value = "task-5106619" [ 872.809867] env[65121]: _type = "Task" [ 872.809867] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.822688] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106619, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.171787] env[65121]: DEBUG nova.network.neutron [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Successfully updated port: 88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 873.189331] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b36e7729-9a68-404a-96ca-b80d564deb32 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "93721179-407a-43d9-b0bf-157433a09519" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.250s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.197773] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106616, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.203728] env[65121]: DEBUG oslo_vmware.api [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.309265] env[65121]: DEBUG nova.compute.manager [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Received event network-changed-33fba103-6efa-42f4-9689-339edc299410 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 873.309500] env[65121]: DEBUG nova.compute.manager [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Refreshing instance network info cache due to event network-changed-33fba103-6efa-42f4-9689-339edc299410. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 873.309742] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Acquiring lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.309888] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Acquired lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.310071] env[65121]: DEBUG nova.network.neutron [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Refreshing network info cache for port 33fba103-6efa-42f4-9689-339edc299410 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 873.311581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-df63596d-e212-4415-8fcf-14df472a0251 tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "b1ad9d15-7ef4-404a-9751-ecdd02ea768b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.260s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.329821] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106619, 'name': ReconfigVM_Task, 'duration_secs': 0.431425} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.329944] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/8a23da27-ce6c-4453-9036-65eeeda3ce0d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.330970] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32912ec7-d88b-425c-a8b2-ab953cac1acc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.340231] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 873.340231] env[65121]: value = "task-5106620" [ 873.340231] env[65121]: _type = "Task" [ 873.340231] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.352631] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106620, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.675558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Acquiring lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.675558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Acquired lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.675869] env[65121]: DEBUG nova.network.neutron [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 873.691772] env[65121]: DEBUG oslo_vmware.api [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106616, 'name': PowerOnVM_Task, 'duration_secs': 1.077382} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.692455] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.692658] env[65121]: INFO nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Took 10.50 seconds to spawn the instance on the hypervisor. [ 873.692830] env[65121]: DEBUG nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 873.693670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0389d3f0-7f57-484d-834a-da1e7a02c8ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.699775] env[65121]: DEBUG nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 873.702278] env[65121]: DEBUG oslo_vmware.api [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106618, 'name': ReconfigVM_Task, 'duration_secs': 0.744464} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.705137] env[65121]: WARNING neutronclient.v2_0.client [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 873.705602] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 873.705674] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Reconfigured VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 873.763469] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.764122] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71f0752e-613b-4fd8-8bc3-c68b52d3c867 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.774796] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 873.774796] env[65121]: value = "task-5106621" [ 873.774796] env[65121]: _type = "Task" [ 873.774796] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.786608] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106621, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.817135] env[65121]: WARNING neutronclient.v2_0.client [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 873.818273] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.819529] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.857980] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106620, 'name': Rename_Task, 'duration_secs': 0.247243} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.858154] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.858480] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a177fef4-fb8a-4893-b778-53534b9e4c3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.871637] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 873.871637] env[65121]: value = "task-5106622" [ 873.871637] env[65121]: _type = "Task" [ 873.871637] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.883072] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106622, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.935031] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.935723] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.994509] env[65121]: WARNING neutronclient.v2_0.client [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 873.994711] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.995110] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.101548] env[65121]: DEBUG nova.network.neutron [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Updated VIF entry in instance network info cache for port 33fba103-6efa-42f4-9689-339edc299410. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 874.102122] env[65121]: DEBUG nova.network.neutron [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Updating instance_info_cache with network_info: [{"id": "33fba103-6efa-42f4-9689-339edc299410", "address": "fa:16:3e:4a:d6:b4", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33fba103-6e", "ovs_interfaceid": "33fba103-6efa-42f4-9689-339edc299410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 874.181318] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.181702] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.212045] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92cad107-f514-42a5-9640-720a068fa315 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-cc002961-d742-4255-88c7-f0a5a39424b1-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.816s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.225460] env[65121]: INFO nova.compute.manager [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Took 57.66 seconds to build instance. [ 874.238212] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.248593] env[65121]: DEBUG nova.network.neutron [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 874.276601] env[65121]: DEBUG nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 874.277171] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 874.277609] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 874.277772] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 874.277932] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 874.278080] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 874.278270] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 874.278409] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.278643] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 874.278930] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 874.279113] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 874.279287] env[65121]: DEBUG nova.virt.hardware [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 874.280868] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70dd9d2-f2ce-4d34-b750-c99a15714697 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.290134] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.290572] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.312697] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa674ee-55fd-47d2-87a2-896767faeea3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.318190] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106621, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.377711] env[65121]: WARNING neutronclient.v2_0.client [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.378542] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.378987] env[65121]: WARNING openstack [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.394929] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106622, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.410334] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13cfc49-f333-4629-8bf3-568f1a08d154 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.418355] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8a6c20-f34a-40f9-a967-76d71b2b33fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.455433] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f993232b-ffa4-4e6a-b798-baa6fed83030 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.464539] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1ccbed-86bb-4e67-845a-7134870a5c2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.479914] env[65121]: DEBUG nova.compute.provider_tree [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.500137] env[65121]: DEBUG nova.network.neutron [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updating instance_info_cache with network_info: [{"id": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "address": "fa:16:3e:e4:16:0d", "network": {"id": "c0b63320-0b56-4ee7-8fa9-654880acb992", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1619842014-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6c6912ada4ba462fb669f0ac353e96ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b9fcb9-e3", "ovs_interfaceid": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 874.606135] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Releasing lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.607063] env[65121]: DEBUG nova.compute.manager [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received event network-vif-plugged-c20059cc-f3b0-4686-acf7-f170edf7da11 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 874.607063] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.607063] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.607063] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.607223] env[65121]: DEBUG nova.compute.manager [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] No waiting events found dispatching network-vif-plugged-c20059cc-f3b0-4686-acf7-f170edf7da11 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 874.607326] env[65121]: WARNING nova.compute.manager [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received unexpected event network-vif-plugged-c20059cc-f3b0-4686-acf7-f170edf7da11 for instance with vm_state active and task_state None. [ 874.607461] env[65121]: DEBUG nova.compute.manager [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received event network-changed-c20059cc-f3b0-4686-acf7-f170edf7da11 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 874.607618] env[65121]: DEBUG nova.compute.manager [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Refreshing instance network info cache due to event network-changed-c20059cc-f3b0-4686-acf7-f170edf7da11. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 874.607790] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Acquiring lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.607917] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Acquired lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.608086] env[65121]: DEBUG nova.network.neutron [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Refreshing network info cache for port c20059cc-f3b0-4686-acf7-f170edf7da11 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 874.726046] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03d9e231-bcba-4edc-810e-d13b1a82f1be tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.655s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.791556] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106621, 'name': PowerOffVM_Task, 'duration_secs': 0.581909} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.791879] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.792147] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.792968] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64541b1-d3e1-49b5-a8bb-f85ad7bd27f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.801422] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.801697] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c519a2d9-12a3-48f0-af65-daf20277f564 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.871916] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.872084] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.872358] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleting the datastore file [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.872850] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72c616d1-b328-499b-850b-f4210ac105d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.887918] env[65121]: DEBUG oslo_vmware.api [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106622, 'name': PowerOnVM_Task, 'duration_secs': 0.544957} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.889448] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.889694] env[65121]: INFO nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Took 6.77 seconds to spawn the instance on the hypervisor. [ 874.889936] env[65121]: DEBUG nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 874.890306] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 874.890306] env[65121]: value = "task-5106624" [ 874.890306] env[65121]: _type = "Task" [ 874.890306] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.891038] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca580e13-b685-412d-aea1-1710728fcf76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.909065] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106624, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.984311] env[65121]: DEBUG nova.scheduler.client.report [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 875.004236] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Releasing lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.004236] env[65121]: DEBUG nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Instance network_info: |[{"id": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "address": "fa:16:3e:e4:16:0d", "network": {"id": "c0b63320-0b56-4ee7-8fa9-654880acb992", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1619842014-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6c6912ada4ba462fb669f0ac353e96ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b9fcb9-e3", "ovs_interfaceid": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 875.005219] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:16:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88b9fcb9-e39c-4751-8499-0f40ea617a1c', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.012312] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Creating folder: Project (6c6912ada4ba462fb669f0ac353e96ed). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.012890] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35d917e8-fa0e-4b8e-ad4e-30678fb57edc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.029548] env[65121]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 875.029548] env[65121]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65121) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 875.029736] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Folder already exists: Project (6c6912ada4ba462fb669f0ac353e96ed). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 875.029962] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Creating folder: Instances. Parent ref: group-v993407. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.030320] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b80b621d-bc81-4d5b-9c55-8cdf75aad8da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.043977] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Created folder: Instances in parent group-v993407. [ 875.044393] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 875.044618] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.045350] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2dc44f24-6bf0-42e3-a291-1f93252bcc84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.072684] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.072684] env[65121]: value = "task-5106627" [ 875.072684] env[65121]: _type = "Task" [ 875.072684] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.082938] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106627, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.111679] env[65121]: WARNING neutronclient.v2_0.client [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 875.112405] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.112922] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.261038] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.261394] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.359859] env[65121]: WARNING neutronclient.v2_0.client [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 875.360781] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.361367] env[65121]: WARNING openstack [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.420881] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106624, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342027} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.421523] env[65121]: INFO nova.compute.manager [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Took 53.49 seconds to build instance. [ 875.428252] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.428453] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.428630] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.490652] env[65121]: DEBUG nova.network.neutron [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updated VIF entry in instance network info cache for port c20059cc-f3b0-4686-acf7-f170edf7da11. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 875.491446] env[65121]: DEBUG nova.network.neutron [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c20059cc-f3b0-4686-acf7-f170edf7da11", "address": "fa:16:3e:07:36:ec", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc20059cc-f3", "ovs_interfaceid": "c20059cc-f3b0-4686-acf7-f170edf7da11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 875.493956] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.751s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 875.494740] env[65121]: DEBUG nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 875.498049] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.919s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.498463] env[65121]: DEBUG nova.objects.instance [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lazy-loading 'resources' on Instance uuid 75bf0b85-db28-4efa-a517-d7256873b09d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.566730] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "c2f9fa49-4273-4885-b59d-0635ed739d65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 875.566980] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.567222] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "c2f9fa49-4273-4885-b59d-0635ed739d65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 875.567400] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.567588] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 875.569961] env[65121]: INFO nova.compute.manager [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Terminating instance [ 875.584144] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106627, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.756947] env[65121]: DEBUG nova.compute.manager [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Received event network-vif-plugged-88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 875.756947] env[65121]: DEBUG oslo_concurrency.lockutils [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Acquiring lock "1b6ad893-e013-4fd5-a829-535d1ea6e001-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 875.756947] env[65121]: DEBUG oslo_concurrency.lockutils [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.756947] env[65121]: DEBUG oslo_concurrency.lockutils [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 875.757299] env[65121]: DEBUG nova.compute.manager [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] No waiting events found dispatching network-vif-plugged-88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 875.757299] env[65121]: WARNING nova.compute.manager [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Received unexpected event network-vif-plugged-88b9fcb9-e39c-4751-8499-0f40ea617a1c for instance with vm_state building and task_state spawning. [ 875.757513] env[65121]: DEBUG nova.compute.manager [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Received event network-changed-88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 875.757513] env[65121]: DEBUG nova.compute.manager [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Refreshing instance network info cache due to event network-changed-88b9fcb9-e39c-4751-8499-0f40ea617a1c. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 875.757676] env[65121]: DEBUG oslo_concurrency.lockutils [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Acquiring lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.757755] env[65121]: DEBUG oslo_concurrency.lockutils [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Acquired lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 875.757932] env[65121]: DEBUG nova.network.neutron [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Refreshing network info cache for port 88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 875.933526] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ec28989a-f2d3-40b4-a60a-73a3ce82bc7c tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.121s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.000263] env[65121]: DEBUG oslo_concurrency.lockutils [req-40ca40ce-98b7-4f08-815d-739a7af2a7e2 req-e174f798-0b89-4c37-9386-3aabebcbf6b7 service nova] Releasing lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 876.002198] env[65121]: DEBUG nova.compute.utils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 876.007277] env[65121]: DEBUG nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 876.075111] env[65121]: DEBUG nova.compute.manager [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 876.075111] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.075844] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d52377-0e7c-479a-93d3-5d2c7c36d71f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.090739] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.094768] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ceec107-a995-4706-9965-b87109bddb8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.096436] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106627, 'name': CreateVM_Task, 'duration_secs': 0.51519} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.098991] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.100187] env[65121]: WARNING neutronclient.v2_0.client [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.100418] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'attachment_id': '1e1db2d6-cbba-4b2e-92e7-e72e30b95fd4', 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993416', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'name': 'volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1b6ad893-e013-4fd5-a829-535d1ea6e001', 'attached_at': '', 'detached_at': '', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'serial': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da'}, 'mount_device': '/dev/sda', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65121) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 876.100633] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Root volume attach. Driver type: vmdk {{(pid=65121) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 876.101793] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250a94b9-d7db-4036-ae1f-2e4c30c3d3ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.107460] env[65121]: DEBUG oslo_vmware.api [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 876.107460] env[65121]: value = "task-5106628" [ 876.107460] env[65121]: _type = "Task" [ 876.107460] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.119310] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad68f93-875e-48d5-9a1c-dede3d92eec8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.127800] env[65121]: DEBUG oslo_vmware.api [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106628, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.140356] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c084ae-01e3-4875-bfb8-230bdde55c07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.152032] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-6127d0d2-2a8c-4bc0-b014-4997be3ef6c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.164748] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 876.164748] env[65121]: value = "task-5106629" [ 876.164748] env[65121]: _type = "Task" [ 876.164748] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.177553] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106629, 'name': RelocateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.196043] env[65121]: INFO nova.compute.manager [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Rebuilding instance [ 876.259605] env[65121]: DEBUG nova.compute.manager [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 876.260416] env[65121]: WARNING neutronclient.v2_0.client [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.261641] env[65121]: WARNING openstack [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.261641] env[65121]: WARNING openstack [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.271073] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0cfdb1-0a0a-405a-befb-1d4f656f9822 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.475196] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 876.475639] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.475912] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 876.476338] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.476742] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 876.476994] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 876.477303] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.477560] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 876.477890] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 876.478209] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 876.478477] env[65121]: DEBUG nova.virt.hardware [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 876.479426] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4e11d3-e18f-4efd-b9e6-955661a586e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.494043] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de45944-9ff0-4acf-bf91-984846e38f11 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.508451] env[65121]: DEBUG nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 876.517617] env[65121]: WARNING openstack [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.517617] env[65121]: WARNING openstack [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.532121] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:da:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2770914-538c-4cb2-9133-0998f766b7ad', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.540040] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 876.543952] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.544849] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb9b7788-bbf4-4e26-bf72-90a9442355b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.572378] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.572378] env[65121]: value = "task-5106630" [ 876.572378] env[65121]: _type = "Task" [ 876.572378] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.589339] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106630, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.619193] env[65121]: DEBUG oslo_vmware.api [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106628, 'name': PowerOffVM_Task, 'duration_secs': 0.283035} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.620450] env[65121]: WARNING neutronclient.v2_0.client [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.621184] env[65121]: WARNING openstack [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.621424] env[65121]: WARNING openstack [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.629620] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.629620] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.632081] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09a4922c-57ff-4d2a-b58f-4782a5f72055 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.664481] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2295d0-b70e-48f7-8915-c56031ff7d5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.680424] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106629, 'name': RelocateVM_Task, 'duration_secs': 0.031947} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.690107] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 876.690432] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993416', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'name': 'volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1b6ad893-e013-4fd5-a829-535d1ea6e001', 'attached_at': '', 'detached_at': '', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'serial': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 876.691960] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5dbd2e-c1b0-4594-9825-b2691d5baa13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.697100] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e2db38-9041-460f-a6bc-4fc8eee91ff5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.765854] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd7288e-ca26-4d47-8be8-5b0db090a17a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.770086] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.770332] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.770507] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleting the datastore file [datastore2] c2f9fa49-4273-4885-b59d-0635ed739d65 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.772773] env[65121]: DEBUG nova.network.neutron [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updated VIF entry in instance network info cache for port 88b9fcb9-e39c-4751-8499-0f40ea617a1c. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 876.773139] env[65121]: DEBUG nova.network.neutron [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updating instance_info_cache with network_info: [{"id": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "address": "fa:16:3e:e4:16:0d", "network": {"id": "c0b63320-0b56-4ee7-8fa9-654880acb992", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1619842014-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6c6912ada4ba462fb669f0ac353e96ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b9fcb9-e3", "ovs_interfaceid": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 876.774912] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e06bb8-b2aa-483b-9169-8ba8fbd9a4bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.777568] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78d643a0-458e-4a30-912c-c971b824118f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.807578] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da/volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.810292] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.810588] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25faa2eb-1898-4e9d-815c-d9f4e3de62a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.826773] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c691636-452d-45b7-9518-b633427165cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.831698] env[65121]: DEBUG oslo_vmware.api [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for the task: (returnval){ [ 876.831698] env[65121]: value = "task-5106632" [ 876.831698] env[65121]: _type = "Task" [ 876.831698] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.831943] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbd3c0a9-4c2d-41d9-abb2-67f06f20e519 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.846979] env[65121]: DEBUG nova.compute.provider_tree [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.851126] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 876.851126] env[65121]: value = "task-5106633" [ 876.851126] env[65121]: _type = "Task" [ 876.851126] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.851829] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 876.851829] env[65121]: value = "task-5106634" [ 876.851829] env[65121]: _type = "Task" [ 876.851829] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.863744] env[65121]: DEBUG oslo_vmware.api [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.870086] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106633, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.875520] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106634, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.008954] env[65121]: INFO nova.compute.manager [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Rescuing [ 877.009280] env[65121]: DEBUG oslo_concurrency.lockutils [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.009465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.009622] env[65121]: DEBUG nova.network.neutron [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 877.088349] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106630, 'name': CreateVM_Task, 'duration_secs': 0.502767} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.088582] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.089166] env[65121]: WARNING neutronclient.v2_0.client [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 877.089548] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.089686] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.090026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 877.090298] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e881555d-6b4f-4d31-bd21-2376accdf219 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.096519] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 877.096519] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526e8273-22a8-b9ff-d2cb-61d86870c3b3" [ 877.096519] env[65121]: _type = "Task" [ 877.096519] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.106370] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526e8273-22a8-b9ff-d2cb-61d86870c3b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.281265] env[65121]: DEBUG oslo_concurrency.lockutils [req-48bc3994-2de8-4db3-ba62-9ff11fbc7ac0 req-bef6eba8-e13e-4e73-a97f-e7a95494ff32 service nova] Releasing lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.347602] env[65121]: DEBUG oslo_vmware.api [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Task: {'id': task-5106632, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247644} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.347602] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.347602] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 877.347602] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 877.347602] env[65121]: INFO nova.compute.manager [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Took 1.27 seconds to destroy the instance on the hypervisor. [ 877.347602] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 877.347602] env[65121]: DEBUG nova.compute.manager [-] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 877.348198] env[65121]: DEBUG nova.network.neutron [-] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 877.348198] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.348465] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.348730] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.356458] env[65121]: DEBUG nova.scheduler.client.report [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 877.375148] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106633, 'name': ReconfigVM_Task, 'duration_secs': 0.388554} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.378797] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Reconfigured VM instance instance-00000041 to attach disk [datastore2] volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da/volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.383753] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106634, 'name': PowerOffVM_Task, 'duration_secs': 0.222318} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.384200] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-cc002961-d742-4255-88c7-f0a5a39424b1-c20059cc-f3b0-4686-acf7-f170edf7da11" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 877.384321] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-cc002961-d742-4255-88c7-f0a5a39424b1-c20059cc-f3b0-4686-acf7-f170edf7da11" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.388086] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2fb51541-8e2b-4eb5-a680-05dddef5481b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.398813] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.399210] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.401323] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e372b127-d2f8-4abc-9db3-9d7398353c78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.406600] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.411213] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.412611] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a780721-3bdc-46fe-a72b-131cdf091af6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.414320] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 877.414320] env[65121]: value = "task-5106635" [ 877.414320] env[65121]: _type = "Task" [ 877.414320] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.424217] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106635, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.502937] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.502937] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.503234] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleting the datastore file [datastore1] 93721179-407a-43d9-b0bf-157433a09519 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.503268] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d46a885-a872-4993-bc8b-3348011adba2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.510276] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 877.510276] env[65121]: value = "task-5106637" [ 877.510276] env[65121]: _type = "Task" [ 877.510276] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.513800] env[65121]: WARNING neutronclient.v2_0.client [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.514865] env[65121]: WARNING openstack [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.515276] env[65121]: WARNING openstack [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.528062] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.541942] env[65121]: DEBUG nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 877.569452] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 877.569685] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 877.571014] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 877.571468] env[65121]: DEBUG nova.virt.hardware [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 877.571962] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0163ac2c-38b7-463f-9c20-271b3a28043a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.585959] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb4798a-8a58-4f36-bc63-54f284852cf7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.607225] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.613243] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Creating folder: Project (da69535027054a7289534ebc2ec26508). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.616873] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-492cd513-60bc-489f-99d4-2578848ae907 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.627203] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526e8273-22a8-b9ff-d2cb-61d86870c3b3, 'name': SearchDatastore_Task, 'duration_secs': 0.012485} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.632780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.632780] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.632780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.632780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.632780] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.632780] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Created folder: Project (da69535027054a7289534ebc2ec26508) in parent group-v993268. [ 877.632780] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Creating folder: Instances. Parent ref: group-v993456. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.632780] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d6761eb7-8884-4ebe-b653-258e030ef00b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.633757] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a72b768-50bb-4bf7-8173-b8e9f1b241e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.649418] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Created folder: Instances in parent group-v993456. [ 877.649625] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 877.649816] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.650038] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2947f0c5-bd90-4a98-a9b9-286680353ee6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.667045] env[65121]: WARNING openstack [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.667630] env[65121]: WARNING openstack [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.680368] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.681267] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.682650] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49c16dbd-5622-457d-bc26-a794ddd2116b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.688764] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.688764] env[65121]: value = "task-5106640" [ 877.688764] env[65121]: _type = "Task" [ 877.688764] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.696055] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 877.696055] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ec0deb-7b64-6b46-ce0d-f1f56dafcf01" [ 877.696055] env[65121]: _type = "Task" [ 877.696055] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.709439] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106640, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.721356] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ec0deb-7b64-6b46-ce0d-f1f56dafcf01, 'name': SearchDatastore_Task, 'duration_secs': 0.017303} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.721824] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1234b699-d3ce-4602-ae6f-1eda80ddea16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.730347] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 877.730347] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52557cc0-2f34-af6b-8f09-77007e92144c" [ 877.730347] env[65121]: _type = "Task" [ 877.730347] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.733823] env[65121]: WARNING neutronclient.v2_0.client [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.734641] env[65121]: WARNING openstack [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.734976] env[65121]: WARNING openstack [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.748495] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52557cc0-2f34-af6b-8f09-77007e92144c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.824042] env[65121]: DEBUG nova.compute.manager [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Received event network-changed-296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 877.824432] env[65121]: DEBUG nova.compute.manager [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Refreshing instance network info cache due to event network-changed-296c4d11-0b03-45a7-8019-6f23e26fea6e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 877.824432] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Acquiring lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.824619] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Acquired lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.824804] env[65121]: DEBUG nova.network.neutron [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Refreshing network info cache for port 296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 877.846052] env[65121]: DEBUG nova.network.neutron [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Updating instance_info_cache with network_info: [{"id": "33fba103-6efa-42f4-9689-339edc299410", "address": "fa:16:3e:4a:d6:b4", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33fba103-6e", "ovs_interfaceid": "33fba103-6efa-42f4-9689-339edc299410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 877.869491] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.371s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.871990] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 44.274s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.872720] env[65121]: DEBUG nova.objects.instance [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lazy-loading 'resources' on Instance uuid ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.891030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.891030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.891030] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f489f7dd-3167-46ce-b15a-69338ff61a86 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.911860] env[65121]: INFO nova.scheduler.client.report [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted allocations for instance 75bf0b85-db28-4efa-a517-d7256873b09d [ 877.917378] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d170352-df4f-4fe2-9483-c8f45022cdd1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.950266] env[65121]: WARNING neutronclient.v2_0.client [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.955424] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Reconfiguring VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 877.959305] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0212d84-7227-43ba-9844-375cfea193fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.973527] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106635, 'name': ReconfigVM_Task, 'duration_secs': 0.21412} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.974592] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993416', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'name': 'volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1b6ad893-e013-4fd5-a829-535d1ea6e001', 'attached_at': '', 'detached_at': '', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'serial': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 877.975675] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e57619d-e9f0-42c9-be73-5e274cdf957e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.984782] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 877.984782] env[65121]: value = "task-5106641" [ 877.984782] env[65121]: _type = "Task" [ 877.984782] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.985951] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 877.985951] env[65121]: value = "task-5106642" [ 877.985951] env[65121]: _type = "Task" [ 877.985951] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.003902] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106642, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.004477] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.022164] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.30442} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.022450] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.022634] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.022805] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.132591] env[65121]: DEBUG nova.network.neutron [-] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 878.201692] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106640, 'name': CreateVM_Task, 'duration_secs': 0.394925} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.201692] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.201692] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.201908] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.202233] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 878.202511] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04352711-c8aa-4643-95dd-acb738acb2c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.208220] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 878.208220] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52326e46-d1ef-e62d-4ab1-9d24fec70e0e" [ 878.208220] env[65121]: _type = "Task" [ 878.208220] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.217700] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52326e46-d1ef-e62d-4ab1-9d24fec70e0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.238557] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52557cc0-2f34-af6b-8f09-77007e92144c, 'name': SearchDatastore_Task, 'duration_secs': 0.018986} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.238827] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.239088] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.239366] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f40ac92-0026-4921-9d09-52076f333859 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.247349] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 878.247349] env[65121]: value = "task-5106643" [ 878.247349] env[65121]: _type = "Task" [ 878.247349] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.257296] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106643, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.327632] env[65121]: WARNING neutronclient.v2_0.client [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.328444] env[65121]: WARNING openstack [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.328884] env[65121]: WARNING openstack [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.349768] env[65121]: DEBUG oslo_concurrency.lockutils [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.442765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e2700dfe-81cb-44f0-a018-e5500cac1773 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "75bf0b85-db28-4efa-a517-d7256873b09d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.523s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.486311] env[65121]: WARNING openstack [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.486587] env[65121]: WARNING openstack [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.516485] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.520955] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106642, 'name': Rename_Task, 'duration_secs': 0.202957} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.527915] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.537727] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb6792a8-1ae3-4f90-b6b9-7617a12f2a37 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.547409] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 878.547409] env[65121]: value = "task-5106644" [ 878.547409] env[65121]: _type = "Task" [ 878.547409] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.560953] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.575063] env[65121]: WARNING neutronclient.v2_0.client [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.575802] env[65121]: WARNING openstack [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.576144] env[65121]: WARNING openstack [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.637176] env[65121]: INFO nova.compute.manager [-] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Took 1.29 seconds to deallocate network for instance. [ 878.676399] env[65121]: DEBUG nova.network.neutron [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Updated VIF entry in instance network info cache for port 296c4d11-0b03-45a7-8019-6f23e26fea6e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 878.676874] env[65121]: DEBUG nova.network.neutron [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Updating instance_info_cache with network_info: [{"id": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "address": "fa:16:3e:b7:7c:fa", "network": {"id": "0004f91f-e9bd-4c04-b81e-2709bf0dbceb", "bridge": "br-int", "label": "tempest-ServersTestJSON-1838786777-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2cae196be0ab4b1ab41f653510f7f87a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "777870ab-362f-4a17-9c1c-8d9cc26cd4ce", "external-id": "nsx-vlan-transportzone-987", "segmentation_id": 987, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap296c4d11-0b", "ovs_interfaceid": "296c4d11-0b03-45a7-8019-6f23e26fea6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 878.719920] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52326e46-d1ef-e62d-4ab1-9d24fec70e0e, 'name': SearchDatastore_Task, 'duration_secs': 0.022059} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.722840] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.723088] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.723406] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.724043] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.724043] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.724242] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c10e2aa-5588-4ba3-8989-fff1bbe6d69e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.738645] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.738831] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.739883] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44dc5d70-29f6-4825-8108-1cbc2f37d032 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.746721] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 878.746721] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52972259-13d6-20c4-0d19-0121d8bcb446" [ 878.746721] env[65121]: _type = "Task" [ 878.746721] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.760392] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52972259-13d6-20c4-0d19-0121d8bcb446, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.769146] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106643, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.012399] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.018367] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c34717d-f87c-4d86-ba28-2eda5a400d10 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.028633] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b0e0fc-44cd-4a93-9e9a-9239f16523fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.071721] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5018403f-3b00-4319-84b1-e108b2f96049 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.082685] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106644, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.087138] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9c7ade-0e36-4b9a-b592-f6e4d8121650 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.105784] env[65121]: DEBUG nova.compute.provider_tree [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.110122] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 879.110122] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 879.110122] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 879.110412] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 879.110618] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 879.110848] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 879.111131] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.111328] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 879.111550] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 879.111764] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 879.111983] env[65121]: DEBUG nova.virt.hardware [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 879.112952] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a15e3a3-0cd0-4980-9ce4-e19e4c3d8ade {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.124743] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d23d51f-3766-4a1c-9696-fb4349865770 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.142468] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:8c:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7651b37-754f-427b-b5a8-5dc1d76d759c', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.151133] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 879.152992] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 879.152992] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.153304] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d598bf75-cdf4-4c6a-96ec-b422e16dca82 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.177371] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.177371] env[65121]: value = "task-5106645" [ 879.177371] env[65121]: _type = "Task" [ 879.177371] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.191622] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] Releasing lock "refresh_cache-7fcab9cb-9a0f-4a50-9244-687889443a31" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.192061] env[65121]: DEBUG nova.compute.manager [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Received event network-vif-deleted-ea99cbd9-05be-40d9-952c-9bee73993946 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 879.192308] env[65121]: INFO nova.compute.manager [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Neutron deleted interface ea99cbd9-05be-40d9-952c-9bee73993946; detaching it from the instance and deleting it from the info cache [ 879.192591] env[65121]: DEBUG nova.network.neutron [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 879.194204] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106645, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.261518] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106643, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.910762} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.265096] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.265519] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.265798] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52972259-13d6-20c4-0d19-0121d8bcb446, 'name': SearchDatastore_Task, 'duration_secs': 0.014099} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.266038] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ae88064-2581-4a43-90a5-6d9ecaf38e81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.269089] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6b0504b-59d3-4305-acc8-6fb9e285f64e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.277467] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 879.277467] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52620ce5-9cde-8f2e-3384-484b15170a80" [ 879.277467] env[65121]: _type = "Task" [ 879.277467] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.279085] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 879.279085] env[65121]: value = "task-5106646" [ 879.279085] env[65121]: _type = "Task" [ 879.279085] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.296388] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52620ce5-9cde-8f2e-3384-484b15170a80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.296714] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.508155] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.578286] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106644, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.618683] env[65121]: DEBUG nova.scheduler.client.report [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 879.644400] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 879.644861] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.689357] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106645, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.695439] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8fb4ccf1-5a5e-4ea2-a3d5-26fb884045ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.707332] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f12dc24-4619-4836-a021-61ef8c0d23f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.749448] env[65121]: DEBUG nova.compute.manager [req-ea16f358-e9c3-45aa-b20e-459bed611ce8 req-28e6f647-0af3-47ce-bbaa-2e781f7a1dde service nova] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Detach interface failed, port_id=ea99cbd9-05be-40d9-952c-9bee73993946, reason: Instance c2f9fa49-4273-4885-b59d-0635ed739d65 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 879.792210] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52620ce5-9cde-8f2e-3384-484b15170a80, 'name': SearchDatastore_Task, 'duration_secs': 0.027744} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.795338] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.795625] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 68e3e4e6-6ca1-4f76-a4af-5112a2042c61/68e3e4e6-6ca1-4f76-a4af-5112a2042c61.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.795901] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074479} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.796122] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5283e6b3-8303-4840-a8c7-25e63b6b956a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.798246] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.799014] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea1d02d-5be3-4fac-be6e-23e4bb3b9c99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.825198] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.827084] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0e38abc-3adc-46c0-929a-5101506fe7cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.841894] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 879.841894] env[65121]: value = "task-5106647" [ 879.841894] env[65121]: _type = "Task" [ 879.841894] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.847371] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 879.847371] env[65121]: value = "task-5106648" [ 879.847371] env[65121]: _type = "Task" [ 879.847371] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.853255] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106647, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.858762] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106648, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.897648] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.897995] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-518595b9-abf0-41d3-85e6-23cdbfe9719a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.907019] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 879.907019] env[65121]: value = "task-5106649" [ 879.907019] env[65121]: _type = "Task" [ 879.907019] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.918214] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.007723] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.079891] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106644, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.124768] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.252s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.128568] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 44.553s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.129432] env[65121]: DEBUG nova.objects.instance [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lazy-loading 'resources' on Instance uuid 90cfd88c-286e-4c87-bf27-909873359997 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.150858] env[65121]: DEBUG nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 880.157490] env[65121]: INFO nova.scheduler.client.report [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Deleted allocations for instance ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7 [ 880.192243] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106645, 'name': CreateVM_Task, 'duration_secs': 0.688232} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.192474] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 880.193150] env[65121]: WARNING neutronclient.v2_0.client [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 880.194028] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.194231] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.194551] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 880.194836] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3418602-a839-429f-b467-b9cbe6ba5900 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.204242] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 880.204242] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525d680c-54e7-09fe-5a7a-c1a230c62c80" [ 880.204242] env[65121]: _type = "Task" [ 880.204242] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.216161] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525d680c-54e7-09fe-5a7a-c1a230c62c80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.355505] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106647, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.363731] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106648, 'name': ReconfigVM_Task, 'duration_secs': 0.483644} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.364027] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.364730] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf59936c-11d8-47f7-aa1f-0369f8f83100 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.375151] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 880.375151] env[65121]: value = "task-5106650" [ 880.375151] env[65121]: _type = "Task" [ 880.375151] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.387332] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106650, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.419120] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106649, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.508925] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.580076] env[65121]: DEBUG oslo_vmware.api [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106644, 'name': PowerOnVM_Task, 'duration_secs': 1.613546} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.580383] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 880.580585] env[65121]: INFO nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Took 6.30 seconds to spawn the instance on the hypervisor. [ 880.580757] env[65121]: DEBUG nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 880.581600] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59663be9-6248-4a58-a4ad-fc9efbe5dd84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.668552] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5c8f9537-77b6-472c-8c4b-2d42084e9a26 tempest-InstanceActionsTestJSON-1092979101 tempest-InstanceActionsTestJSON-1092979101-project-member] Lock "ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 50.630s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.675411] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.717978] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525d680c-54e7-09fe-5a7a-c1a230c62c80, 'name': SearchDatastore_Task, 'duration_secs': 0.058171} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.718335] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 880.718581] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.718822] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.719026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.719169] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.719457] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-670bafe0-12d2-4eea-af02-11a295ac30de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.733905] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.734124] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.737959] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-874b031e-70bf-4d0e-adde-c24b5542482a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.747244] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 880.747244] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5200f6bb-f726-222b-d314-20ec5039eee8" [ 880.747244] env[65121]: _type = "Task" [ 880.747244] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.756106] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5200f6bb-f726-222b-d314-20ec5039eee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.851919] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106647, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688807} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.854485] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 68e3e4e6-6ca1-4f76-a4af-5112a2042c61/68e3e4e6-6ca1-4f76-a4af-5112a2042c61.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.854724] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.855363] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c3ead009-567d-4264-a3e1-b1741daf8148 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.862807] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 880.862807] env[65121]: value = "task-5106651" [ 880.862807] env[65121]: _type = "Task" [ 880.862807] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.873137] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106651, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.888160] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106650, 'name': Rename_Task, 'duration_secs': 0.167797} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.888446] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.888702] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5478b23e-869a-4edf-a9b1-cfce53baf87a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.895416] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 880.895416] env[65121]: value = "task-5106652" [ 880.895416] env[65121]: _type = "Task" [ 880.895416] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.907763] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.919523] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106649, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.011886] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.104021] env[65121]: INFO nova.compute.manager [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Took 53.57 seconds to build instance. [ 881.115690] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234776a2-0316-4ee3-b636-6a35e8a7ef4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.124545] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03907741-cc18-41dd-8c18-1282a63b7ceb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.166168] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1432799-a54f-4cf0-be07-d1cde0984590 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.177708] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5829d545-b943-42c0-85d9-8feaeefc1637 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.198732] env[65121]: DEBUG nova.compute.provider_tree [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.262062] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5200f6bb-f726-222b-d314-20ec5039eee8, 'name': SearchDatastore_Task, 'duration_secs': 0.042614} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.262062] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db543273-7bb6-4d34-bb78-961830e8e56e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.268966] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 881.268966] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529722bd-6d9c-1b03-33ac-f25e46648be2" [ 881.268966] env[65121]: _type = "Task" [ 881.268966] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.279937] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529722bd-6d9c-1b03-33ac-f25e46648be2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.372784] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106651, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07669} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.372922] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.374104] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdd8b6f-a73a-480a-8156-d459c1ae613b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.395493] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 68e3e4e6-6ca1-4f76-a4af-5112a2042c61/68e3e4e6-6ca1-4f76-a4af-5112a2042c61.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.395822] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c06f9507-50ba-4eb7-a213-6ad28973be46 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.423406] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106652, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.427764] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106649, 'name': PowerOffVM_Task, 'duration_secs': 1.21622} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.427925] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 881.427925] env[65121]: value = "task-5106653" [ 881.427925] env[65121]: _type = "Task" [ 881.427925] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.431441] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.431441] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6a3966-208b-4b3c-9527-fa8bad5ad4c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.443927] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.459388] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa9b44b-8508-4ef9-9506-9c413741895c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.501828] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.502337] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f557eabc-05ac-45ce-b236-dab7e31f847c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.516554] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.518401] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 881.518401] env[65121]: value = "task-5106654" [ 881.518401] env[65121]: _type = "Task" [ 881.518401] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.528953] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 881.529238] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.529622] env[65121]: DEBUG oslo_concurrency.lockutils [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.529719] env[65121]: DEBUG oslo_concurrency.lockutils [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.529913] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.530235] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-10f37da7-c085-45b5-aec0-739a0e7f85c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.544646] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.544881] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.545681] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60662608-61de-458a-b7e0-4a3f919f85a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.552339] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 881.552339] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5299adfd-dc82-d3e4-c2dd-de42c10aac14" [ 881.552339] env[65121]: _type = "Task" [ 881.552339] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.561299] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5299adfd-dc82-d3e4-c2dd-de42c10aac14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.606205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3f19da35-dd69-4b1d-8572-212da058a4d5 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.075s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 881.703034] env[65121]: DEBUG nova.scheduler.client.report [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 881.782754] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529722bd-6d9c-1b03-33ac-f25e46648be2, 'name': SearchDatastore_Task, 'duration_secs': 0.015642} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.782754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.782754] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.782754] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05850e97-ace7-41d7-ba0e-d0c48c82a1f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.789779] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 881.789779] env[65121]: value = "task-5106655" [ 881.789779] env[65121]: _type = "Task" [ 881.789779] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.799502] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.920834] env[65121]: DEBUG oslo_vmware.api [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106652, 'name': PowerOnVM_Task, 'duration_secs': 0.710996} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.921134] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.921342] env[65121]: DEBUG nova.compute.manager [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 881.922185] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2270401a-b866-469a-95fa-71bad7cd32cc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.944592] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106653, 'name': ReconfigVM_Task, 'duration_secs': 0.304418} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.944987] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 68e3e4e6-6ca1-4f76-a4af-5112a2042c61/68e3e4e6-6ca1-4f76-a4af-5112a2042c61.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.945807] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5985af8c-f989-4bf3-a0b7-8e02946c8466 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.954540] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 881.954540] env[65121]: value = "task-5106656" [ 881.954540] env[65121]: _type = "Task" [ 881.954540] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.968925] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106656, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.013603] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.069388] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5299adfd-dc82-d3e4-c2dd-de42c10aac14, 'name': SearchDatastore_Task, 'duration_secs': 0.019771} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.070673] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1652879-ff36-4611-9e06-042d494627dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.079377] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 882.079377] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e82394-edcf-9851-5302-5c93cb07d8be" [ 882.079377] env[65121]: _type = "Task" [ 882.079377] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.089488] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e82394-edcf-9851-5302-5c93cb07d8be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.209357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.081s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.211770] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.130s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.213446] env[65121]: INFO nova.compute.claims [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.233483] env[65121]: INFO nova.scheduler.client.report [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleted allocations for instance 90cfd88c-286e-4c87-bf27-909873359997 [ 882.311585] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106655, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.445379] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.472644] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106656, 'name': Rename_Task, 'duration_secs': 0.204326} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.473087] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.473389] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b89b5d50-8103-45b9-9b81-1555f7cb6424 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.484800] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 882.484800] env[65121]: value = "task-5106657" [ 882.484800] env[65121]: _type = "Task" [ 882.484800] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.497824] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106657, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.521350] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.596522] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e82394-edcf-9851-5302-5c93cb07d8be, 'name': SearchDatastore_Task, 'duration_secs': 0.016619} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.596522] env[65121]: DEBUG oslo_concurrency.lockutils [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.596522] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. {{(pid=65121) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 882.596522] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b45ef9a9-48f7-40ef-8dd3-44dee19936ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.606451] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 882.606451] env[65121]: value = "task-5106658" [ 882.606451] env[65121]: _type = "Task" [ 882.606451] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.618712] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.741925] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b0900a25-f167-4626-b35e-1ebb5e720b90 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "90cfd88c-286e-4c87-bf27-909873359997" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 50.632s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.807433] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106655, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.786906} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.807737] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.807973] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.808252] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b7b6cbe-c6dc-44b1-8b73-467b296c99f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.820386] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 882.820386] env[65121]: value = "task-5106659" [ 882.820386] env[65121]: _type = "Task" [ 882.820386] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.833326] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106659, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.003957] env[65121]: DEBUG oslo_vmware.api [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106657, 'name': PowerOnVM_Task, 'duration_secs': 0.490149} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.004448] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.004632] env[65121]: INFO nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Took 5.46 seconds to spawn the instance on the hypervisor. [ 883.004812] env[65121]: DEBUG nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 883.015079] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712c0505-f1c7-4c4e-95f5-2f8d35a4d67d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.030538] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.041695] env[65121]: DEBUG nova.compute.manager [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Received event network-changed-88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 883.041994] env[65121]: DEBUG nova.compute.manager [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Refreshing instance network info cache due to event network-changed-88b9fcb9-e39c-4751-8499-0f40ea617a1c. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 883.042374] env[65121]: DEBUG oslo_concurrency.lockutils [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Acquiring lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.042609] env[65121]: DEBUG oslo_concurrency.lockutils [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Acquired lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.043222] env[65121]: DEBUG nova.network.neutron [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Refreshing network info cache for port 88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 883.120535] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106658, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.332565] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106659, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.162632} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.332865] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.333986] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a2f070-7e7f-40a6-8e14-c30b26b1e4c5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.366577] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.366915] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f7416f1-3592-46ba-9ddc-2f0826dbe06a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.395590] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 883.395590] env[65121]: value = "task-5106660" [ 883.395590] env[65121]: _type = "Task" [ 883.395590] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.405258] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106660, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.516165] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.547374] env[65121]: WARNING neutronclient.v2_0.client [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.548097] env[65121]: WARNING openstack [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.548429] env[65121]: WARNING openstack [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.562198] env[65121]: INFO nova.compute.manager [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Took 52.58 seconds to build instance. [ 883.621216] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106658, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607111} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.621484] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. [ 883.622310] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85a8524-90f0-4541-9b3c-ce9a87b27a61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.652831] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.655842] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db18889a-0efb-4091-b567-bd8cf39ae27e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.682653] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 883.682653] env[65121]: value = "task-5106661" [ 883.682653] env[65121]: _type = "Task" [ 883.682653] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.694498] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.724745] env[65121]: WARNING openstack [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.725219] env[65121]: WARNING openstack [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.817781] env[65121]: WARNING neutronclient.v2_0.client [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.818587] env[65121]: WARNING openstack [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.819753] env[65121]: WARNING openstack [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.880170] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a7eeff-fc86-4c40-95a2-8b91e44127c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.903072] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d0a72d-6b41-479e-ab2a-aa68dcb19ea8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.911130] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106660, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.948480] env[65121]: DEBUG nova.network.neutron [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updated VIF entry in instance network info cache for port 88b9fcb9-e39c-4751-8499-0f40ea617a1c. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 883.948886] env[65121]: DEBUG nova.network.neutron [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updating instance_info_cache with network_info: [{"id": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "address": "fa:16:3e:e4:16:0d", "network": {"id": "c0b63320-0b56-4ee7-8fa9-654880acb992", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1619842014-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6c6912ada4ba462fb669f0ac353e96ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b9fcb9-e3", "ovs_interfaceid": "88b9fcb9-e39c-4751-8499-0f40ea617a1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 883.953772] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e62a36b-5af9-40cb-8eb7-3868016b6391 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.964236] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1709742c-478c-41e8-a25b-a7f0887e500c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.985642] env[65121]: DEBUG nova.compute.provider_tree [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.017184] env[65121]: DEBUG oslo_vmware.api [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106641, 'name': ReconfigVM_Task, 'duration_secs': 5.981199} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.017900] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.017900] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Reconfigured VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 884.018220] env[65121]: WARNING neutronclient.v2_0.client [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.019072] env[65121]: WARNING neutronclient.v2_0.client [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.019186] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.019970] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.064628] env[65121]: DEBUG oslo_concurrency.lockutils [None req-984a4286-20af-493c-a403-9898639b2ff9 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.214s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 884.195292] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106661, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.409082] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106660, 'name': ReconfigVM_Task, 'duration_secs': 0.930336} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.409082] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 93721179-407a-43d9-b0bf-157433a09519/93721179-407a-43d9-b0bf-157433a09519.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.410121] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f67f4c48-2b3c-4dfe-b490-b3cd87b775a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.418586] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 884.418586] env[65121]: value = "task-5106662" [ 884.418586] env[65121]: _type = "Task" [ 884.418586] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.431735] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106662, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.459299] env[65121]: DEBUG oslo_concurrency.lockutils [req-994e0593-6311-410f-9343-c88853bdac3e req-d75f43a4-3f76-45a2-a310-9e4176c52779 service nova] Releasing lock "refresh_cache-1b6ad893-e013-4fd5-a829-535d1ea6e001" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.460146] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "49b7e6c3-9106-4b57-ae44-85b5d4474089" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.460530] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.489434] env[65121]: DEBUG nova.scheduler.client.report [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 884.693345] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106661, 'name': ReconfigVM_Task, 'duration_secs': 0.697037} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.693627] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.694516] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7d14d4-b98c-440d-9207-ec81bb19c809 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.723590] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e33777a0-44e7-49e1-80c9-5b8fe4d77965 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.741797] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 884.741797] env[65121]: value = "task-5106663" [ 884.741797] env[65121]: _type = "Task" [ 884.741797] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.752850] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106663, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.930047] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106662, 'name': Rename_Task, 'duration_secs': 0.226034} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.930382] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.930699] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30f78860-7364-491e-8751-05feb3af4583 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.938708] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 884.938708] env[65121]: value = "task-5106664" [ 884.938708] env[65121]: _type = "Task" [ 884.938708] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.948535] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.962787] env[65121]: DEBUG nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 884.998530] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.786s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 884.998530] env[65121]: DEBUG nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 885.002285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 42.400s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.141484] env[65121]: INFO nova.compute.manager [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Rebuilding instance [ 885.148081] env[65121]: DEBUG nova.compute.manager [req-108480ae-867e-4a5d-a5f9-78132fed62ad req-ff0b1bc6-6065-4924-a857-061dc5585512 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received event network-vif-deleted-c20059cc-f3b0-4686-acf7-f170edf7da11 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 885.148353] env[65121]: INFO nova.compute.manager [req-108480ae-867e-4a5d-a5f9-78132fed62ad req-ff0b1bc6-6065-4924-a857-061dc5585512 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Neutron deleted interface c20059cc-f3b0-4686-acf7-f170edf7da11; detaching it from the instance and deleting it from the info cache [ 885.148580] env[65121]: DEBUG nova.network.neutron [req-108480ae-867e-4a5d-a5f9-78132fed62ad req-ff0b1bc6-6065-4924-a857-061dc5585512 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 885.210770] env[65121]: DEBUG nova.compute.manager [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 885.211688] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c034272-d164-4757-b259-d07fcefe3184 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.252369] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106663, 'name': ReconfigVM_Task, 'duration_secs': 0.428453} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.252690] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.253069] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46a63501-6667-4790-a587-1dcac376cb14 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.261324] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 885.261324] env[65121]: value = "task-5106665" [ 885.261324] env[65121]: _type = "Task" [ 885.261324] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.271367] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106665, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.408686] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.408897] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.409064] env[65121]: DEBUG nova.network.neutron [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 885.452398] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106664, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.488681] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.505941] env[65121]: DEBUG nova.compute.utils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 885.520231] env[65121]: DEBUG nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 885.520439] env[65121]: DEBUG nova.network.neutron [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 885.520766] env[65121]: WARNING neutronclient.v2_0.client [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.521073] env[65121]: WARNING neutronclient.v2_0.client [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.521749] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.522111] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.533797] env[65121]: DEBUG nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 885.651837] env[65121]: DEBUG oslo_concurrency.lockutils [req-108480ae-867e-4a5d-a5f9-78132fed62ad req-ff0b1bc6-6065-4924-a857-061dc5585512 service nova] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.652080] env[65121]: DEBUG oslo_concurrency.lockutils [req-108480ae-867e-4a5d-a5f9-78132fed62ad req-ff0b1bc6-6065-4924-a857-061dc5585512 service nova] Acquired lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.653046] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0a9759-b4e0-4a6e-ab26-1b7edc037132 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.675811] env[65121]: DEBUG oslo_concurrency.lockutils [req-108480ae-867e-4a5d-a5f9-78132fed62ad req-ff0b1bc6-6065-4924-a857-061dc5585512 service nova] Releasing lock "cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.676184] env[65121]: WARNING nova.compute.manager [req-108480ae-867e-4a5d-a5f9-78132fed62ad req-ff0b1bc6-6065-4924-a857-061dc5585512 service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Detach interface failed, port_id=c20059cc-f3b0-4686-acf7-f170edf7da11, reason: No device with interface-id c20059cc-f3b0-4686-acf7-f170edf7da11 exists on VM: nova.exception.NotFound: No device with interface-id c20059cc-f3b0-4686-acf7-f170edf7da11 exists on VM [ 885.710335] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.710335] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "cc002961-d742-4255-88c7-f0a5a39424b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.710658] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.711042] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.711165] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "cc002961-d742-4255-88c7-f0a5a39424b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.714444] env[65121]: INFO nova.compute.manager [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Terminating instance [ 885.773599] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106665, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.913802] env[65121]: WARNING neutronclient.v2_0.client [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.914703] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.914944] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.950623] env[65121]: DEBUG oslo_vmware.api [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106664, 'name': PowerOnVM_Task, 'duration_secs': 0.807275} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.950962] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.951191] env[65121]: DEBUG nova.compute.manager [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 885.952100] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbf6b95-7d74-4934-8c4c-018b8afba452 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.024020] env[65121]: DEBUG nova.policy [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af00bd582c0843949491bbcecbfcd2dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72237a0d762645588c41231b0a34a796', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 886.034498] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Applying migration context for instance 293f93f2-c01d-42c8-b1a7-3056805c77de as it has an incoming, in-progress migration 2ce31b97-0aef-4465-995a-13b22493087b. Migration status is reverting {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 886.038092] env[65121]: INFO nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating resource usage from migration 2ce31b97-0aef-4465-995a-13b22493087b [ 886.038830] env[65121]: INFO nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating resource usage from migration 71f1858e-40ac-4394-b0fc-2d77e0195ba7 [ 886.060527] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance cc002961-d742-4255-88c7-f0a5a39424b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.060693] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f1a5fdce-a940-43d4-83d7-716786c9ac34 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.060820] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c4096314-270a-4270-9e1d-5ace8ddbd286 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.060940] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 5f6e4f46-0745-42c0-a779-6cffb60e21ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.061069] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c993d5a4-49d6-43aa-a1f0-4aac91fad953 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.061196] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c2f9fa49-4273-4885-b59d-0635ed739d65 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.061562] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 27940143-16b5-4263-b23c-354ed8ea8866 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.061562] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 9cf6c29a-3424-4b88-9ba5-8120b124beb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.061562] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f84d7aa1-a90b-4c95-9673-fb1a2af7187c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.061666] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance b6f6871d-311c-4adb-824e-2907a12f4224 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.061699] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance ab41ae41-e69d-47fb-a31a-16748fd3b0f2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.061793] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4a72fedd-b114-468e-8f34-0caec6ce73fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.061893] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 3f3bec5b-2834-497c-a454-a152cb992309 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.062017] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance e2e03e5c-4a71-4555-9cd4-d67e178064ec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.062617] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance e0d3e3c8-6471-4345-8677-369612674769 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.062617] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance b22ccce7-c54a-4577-9de0-1fd9c10cd189 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.062617] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Migration 2ce31b97-0aef-4465-995a-13b22493087b is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 886.062617] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 293f93f2-c01d-42c8-b1a7-3056805c77de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.062617] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 5d352184-7074-4fab-b4ac-cc11a6b936b1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.062811] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 99b1ff9c-38b7-4947-81bb-d200a3b2a09a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.062811] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 93721179-407a-43d9-b0bf-157433a09519 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.062875] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 7fcab9cb-9a0f-4a50-9244-687889443a31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.062950] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 8a23da27-ce6c-4453-9036-65eeeda3ce0d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.063239] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 1b6ad893-e013-4fd5-a829-535d1ea6e001 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.063412] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 68e3e4e6-6ca1-4f76-a4af-5112a2042c61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.063875] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 274d0ccd-c707-4a68-b280-16de2bc74d73 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 886.219927] env[65121]: DEBUG nova.compute.manager [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 886.220185] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.221905] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29d5028-3e58-4baf-94db-99f4c06e248f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.227138] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.227138] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbb6ae29-ed4e-4129-8fda-1b221e4a54e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.235263] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.236633] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98f2d528-8cea-403f-ac24-79e6acb705bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.238306] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 886.238306] env[65121]: value = "task-5106666" [ 886.238306] env[65121]: _type = "Task" [ 886.238306] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.245059] env[65121]: DEBUG oslo_vmware.api [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 886.245059] env[65121]: value = "task-5106667" [ 886.245059] env[65121]: _type = "Task" [ 886.245059] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.254496] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106666, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.263681] env[65121]: DEBUG oslo_vmware.api [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106667, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.272922] env[65121]: DEBUG oslo_vmware.api [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106665, 'name': PowerOnVM_Task, 'duration_secs': 0.691279} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.273209] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.277042] env[65121]: DEBUG nova.compute.manager [None req-24875fa0-b278-4d6e-9983-9c8f9f784f50 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 886.277911] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92617abe-3f33-4697-a453-7a6c5c4b53ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.482016] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 886.519605] env[65121]: DEBUG nova.network.neutron [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Successfully created port: 748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 886.549231] env[65121]: DEBUG nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 886.568232] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 886.582035] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 886.582035] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 886.582035] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 886.582035] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 886.582428] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 886.582428] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 886.582534] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.582579] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 886.582744] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 886.582894] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 886.583147] env[65121]: DEBUG nova.virt.hardware [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 886.584191] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6f6ce4-dc61-43ab-8e3e-b83d3713dcbd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.596509] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e20b81-6656-498b-b535-53ca713103d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.648358] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.648862] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.754679] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106666, 'name': PowerOffVM_Task, 'duration_secs': 0.436622} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.755095] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.756110] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.759104] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856f7eef-dd50-4488-9422-c0a52efd6159 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.772605] env[65121]: DEBUG oslo_vmware.api [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106667, 'name': PowerOffVM_Task, 'duration_secs': 0.426321} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.773188] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.773188] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.773311] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.773511] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1998fd75-1376-4484-b13b-fc7c8b68e53c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.775136] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f8d90cd-6ddd-4c68-837d-684af4b8bb2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.777489] env[65121]: WARNING neutronclient.v2_0.client [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.778114] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.778441] env[65121]: WARNING openstack [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.983522] env[65121]: DEBUG nova.network.neutron [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [{"id": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "address": "fa:16:3e:c7:ed:b2", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13f8bd13-f8", "ovs_interfaceid": "13f8bd13-f8f0-4818-8dd6-845e9644a395", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 887.074877] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance df5abcb5-583f-4b28-a074-3a3221d74d87 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 887.328333] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.328635] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.329455] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleting the datastore file [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.329698] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b6d1e84-9740-4eee-98a1-0c4324b30558 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.331907] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.332107] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.332298] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleting the datastore file [datastore1] cc002961-d742-4255-88c7-f0a5a39424b1 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.332569] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-324ec47f-5125-436f-b721-bb09091b3646 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.347513] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 887.347513] env[65121]: value = "task-5106670" [ 887.347513] env[65121]: _type = "Task" [ 887.347513] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.347513] env[65121]: DEBUG oslo_vmware.api [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 887.347513] env[65121]: value = "task-5106671" [ 887.347513] env[65121]: _type = "Task" [ 887.347513] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.364699] env[65121]: DEBUG oslo_vmware.api [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106671, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.368706] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.485999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-cc002961-d742-4255-88c7-f0a5a39424b1" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.578559] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 2e676b64-181e-4fee-8120-05cb49b5bb4e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 887.861973] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180884} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.863197] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.863422] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.863604] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.869954] env[65121]: DEBUG oslo_vmware.api [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106671, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322535} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.870449] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.870659] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.871594] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.871815] env[65121]: INFO nova.compute.manager [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Took 1.65 seconds to destroy the instance on the hypervisor. [ 887.872085] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 887.872286] env[65121]: DEBUG nova.compute.manager [-] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 887.872383] env[65121]: DEBUG nova.network.neutron [-] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 887.872638] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.873227] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.873493] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.916841] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.994754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15b7a7f3-8057-47b8-9539-7bb92db127cc tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-cc002961-d742-4255-88c7-f0a5a39424b1-c20059cc-f3b0-4686-acf7-f170edf7da11" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.609s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.082417] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 600d5554-f52a-48ca-941d-1a755d086823 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 888.296932] env[65121]: INFO nova.compute.manager [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Unrescuing [ 888.297536] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.297644] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquired lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.297938] env[65121]: DEBUG nova.network.neutron [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 888.313638] env[65121]: DEBUG nova.network.neutron [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Successfully updated port: 748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 888.566182] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "ac001ee6-f601-48f7-af0d-42c2387f3c98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 888.566182] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.586403] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance b2ebce1d-a8d8-4a71-997b-39d09ca373fe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 888.586638] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Migration 71f1858e-40ac-4394-b0fc-2d77e0195ba7 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 888.586826] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 888.776254] env[65121]: DEBUG nova.compute.manager [req-de463d49-2068-4fdd-9ad5-ae350930d7b4 req-a8c4b3dd-d3a4-4ed1-b3fe-d6e27353c225 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Received event network-vif-plugged-748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 888.777236] env[65121]: DEBUG oslo_concurrency.lockutils [req-de463d49-2068-4fdd-9ad5-ae350930d7b4 req-a8c4b3dd-d3a4-4ed1-b3fe-d6e27353c225 service nova] Acquiring lock "274d0ccd-c707-4a68-b280-16de2bc74d73-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 888.777236] env[65121]: DEBUG oslo_concurrency.lockutils [req-de463d49-2068-4fdd-9ad5-ae350930d7b4 req-a8c4b3dd-d3a4-4ed1-b3fe-d6e27353c225 service nova] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.777236] env[65121]: DEBUG oslo_concurrency.lockutils [req-de463d49-2068-4fdd-9ad5-ae350930d7b4 req-a8c4b3dd-d3a4-4ed1-b3fe-d6e27353c225 service nova] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.777236] env[65121]: DEBUG nova.compute.manager [req-de463d49-2068-4fdd-9ad5-ae350930d7b4 req-a8c4b3dd-d3a4-4ed1-b3fe-d6e27353c225 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] No waiting events found dispatching network-vif-plugged-748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 888.777236] env[65121]: WARNING nova.compute.manager [req-de463d49-2068-4fdd-9ad5-ae350930d7b4 req-a8c4b3dd-d3a4-4ed1-b3fe-d6e27353c225 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Received unexpected event network-vif-plugged-748bfc22-93f1-459b-9b59-18583587dd17 for instance with vm_state building and task_state spawning. [ 888.800940] env[65121]: WARNING neutronclient.v2_0.client [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.801607] env[65121]: WARNING openstack [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.802674] env[65121]: WARNING openstack [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.817438] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.817626] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.818282] env[65121]: DEBUG nova.network.neutron [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 888.918496] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 888.918800] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 888.918972] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 888.919908] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 888.920351] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 888.920515] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 888.920732] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.920887] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 888.921079] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 888.921239] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 888.921409] env[65121]: DEBUG nova.virt.hardware [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 888.922350] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b2b62e-fdc2-47e0-8412-1914694abcae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.930909] env[65121]: WARNING openstack [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.931301] env[65121]: WARNING openstack [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.950700] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7dc21b5-10ab-4815-bd38-0d06e42f89f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.969112] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:da:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2770914-538c-4cb2-9133-0998f766b7ad', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.977840] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 888.983414] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.984297] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7bef5a88-68e9-459d-b53e-f12c0b787310 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.001237] env[65121]: WARNING neutronclient.v2_0.client [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 889.001888] env[65121]: WARNING openstack [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.002240] env[65121]: WARNING openstack [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.013269] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.013269] env[65121]: value = "task-5106672" [ 889.013269] env[65121]: _type = "Task" [ 889.013269] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.023093] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106672, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.040031] env[65121]: DEBUG nova.network.neutron [-] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.068446] env[65121]: DEBUG nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 889.090501] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c52a1269-bb34-4ef6-ab8d-78df3e9b1d39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 889.109021] env[65121]: DEBUG nova.network.neutron [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Updating instance_info_cache with network_info: [{"id": "33fba103-6efa-42f4-9689-339edc299410", "address": "fa:16:3e:4a:d6:b4", "network": {"id": "af61c465-54f5-43c2-b90a-59441c766a4c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-2120966458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "693c1c83a1d5453bbedaf8ce2bf7c8d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33fba103-6e", "ovs_interfaceid": "33fba103-6efa-42f4-9689-339edc299410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.164692] env[65121]: DEBUG nova.compute.manager [req-3a521645-fbf1-49fc-af7e-8fcf6f0b1c82 req-74da5b1f-05a9-4987-b302-469586ed195e service nova] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Received event network-vif-deleted-13f8bd13-f8f0-4818-8dd6-845e9644a395 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 889.322970] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.324026] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.362163] env[65121]: DEBUG nova.network.neutron [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 889.384442] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.384442] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.454446] env[65121]: WARNING neutronclient.v2_0.client [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 889.455105] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.455475] env[65121]: WARNING openstack [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.523954] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106672, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.546383] env[65121]: INFO nova.compute.manager [-] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Took 1.67 seconds to deallocate network for instance. [ 889.549044] env[65121]: DEBUG nova.network.neutron [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Updating instance_info_cache with network_info: [{"id": "748bfc22-93f1-459b-9b59-18583587dd17", "address": "fa:16:3e:f4:04:0c", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap748bfc22-93", "ovs_interfaceid": "748bfc22-93f1-459b-9b59-18583587dd17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.551859] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "93721179-407a-43d9-b0bf-157433a09519" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.552031] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "93721179-407a-43d9-b0bf-157433a09519" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.552217] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "93721179-407a-43d9-b0bf-157433a09519-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.552999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "93721179-407a-43d9-b0bf-157433a09519-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.552999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "93721179-407a-43d9-b0bf-157433a09519-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.560343] env[65121]: INFO nova.compute.manager [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Terminating instance [ 889.594459] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 49b7e6c3-9106-4b57-ae44-85b5d4474089 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 889.595301] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 22 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 889.595301] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4864MB phys_disk=100GB used_disk=22GB total_vcpus=48 used_vcpus=22 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '21', 'num_vm_active': '17', 'num_task_None': '16', 'num_os_type_None': '21', 'num_proj_982f66705583488cb369f19160cc2ee5': '2', 'io_workload': '4', 'num_proj_35312a302644426f98f127e89a067e75': '1', 'num_proj_72237a0d762645588c41231b0a34a796': '2', 'num_task_rebuilding': '1', 'num_proj_083c4f2aa7bf4167b472db6bebb9d8bd': '4', 'num_vm_stopped': '1', 'num_task_resize_prep': '1', 'num_proj_4467da3ed41245ddbc93fc865a8b7bdd': '1', 'num_proj_89110618b6ef4ce68b2ca53e7e168139': '1', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_vm_rescued': '1', 'num_proj_693c1c83a1d5453bbedaf8ce2bf7c8d1': '2', 'num_proj_00383aa3355e438cb703c2b86c7917f2': '1', 'num_proj_40aee06073b04c25bc7b38fbd09856e9': '1', 'num_proj_ad1aea30d62c45e193c5a54c429ce7ba': '1', 'num_proj_741e03af2f4d451d9ed77004b644deb8': '1', 'num_task_rebuild_spawning': '1', 'num_proj_e491b88dcaad42dcb345b3217bc6f941': '1', 'num_proj_2cae196be0ab4b1ab41f653510f7f87a': '1', 'num_task_rescuing': '1', 'num_proj_6c6912ada4ba462fb669f0ac353e96ed': '1', 'num_proj_da69535027054a7289534ebc2ec26508': '1', 'num_vm_building': '1'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 889.599327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.611011] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Releasing lock "refresh_cache-8a23da27-ce6c-4453-9036-65eeeda3ce0d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.614774] env[65121]: DEBUG nova.objects.instance [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lazy-loading 'flavor' on Instance uuid 8a23da27-ce6c-4453-9036-65eeeda3ce0d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.024803] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106672, 'name': CreateVM_Task, 'duration_secs': 0.749938} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.024970] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.025667] env[65121]: WARNING neutronclient.v2_0.client [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 890.025813] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.025967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.026267] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 890.028969] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08d50625-f844-4f50-9924-39c90b43ebda {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.036156] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 890.036156] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52491751-c7cd-b462-5538-f68811def181" [ 890.036156] env[65121]: _type = "Task" [ 890.036156] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.043996] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52491751-c7cd-b462-5538-f68811def181, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.058912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.059296] env[65121]: DEBUG nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Instance network_info: |[{"id": "748bfc22-93f1-459b-9b59-18583587dd17", "address": "fa:16:3e:f4:04:0c", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap748bfc22-93", "ovs_interfaceid": "748bfc22-93f1-459b-9b59-18583587dd17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 890.060657] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.061284] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:04:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6966f473-59ac-49bb-9b7a-22c61f4e61e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '748bfc22-93f1-459b-9b59-18583587dd17', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.069280] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 890.069607] env[65121]: DEBUG nova.compute.manager [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 890.070099] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 890.070099] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.070862] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70793228-3656-48d8-b745-896b8a94fc18 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.076106] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13a47c70-630a-4481-b681-7db1f8f7b595 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.096312] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.097751] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96595928-c5aa-46c1-8cae-9b51e2ad02dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.099454] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.099454] env[65121]: value = "task-5106673" [ 890.099454] env[65121]: _type = "Task" [ 890.099454] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.107022] env[65121]: DEBUG oslo_vmware.api [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 890.107022] env[65121]: value = "task-5106674" [ 890.107022] env[65121]: _type = "Task" [ 890.107022] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.113296] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106673, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.121501] env[65121]: DEBUG oslo_vmware.api [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.125466] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c779d0f2-f2d1-4615-9a1c-5e9de25f277f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.153698] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.157340] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21d2465d-0919-4005-ba47-031095f71762 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.165836] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 890.165836] env[65121]: value = "task-5106675" [ 890.165836] env[65121]: _type = "Task" [ 890.165836] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.176957] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106675, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.180630] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff03a50-629d-4c25-a3bd-8dfe30c2a4c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.189714] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015d9879-628c-427f-9e19-2e5cb59682a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.229257] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8132c95-d82f-4203-b7d7-2d1cf54de2e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.240579] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34beeaba-5cba-4cc9-96ea-7a87e788ef7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.257656] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.341510] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "59d6005d-6caf-4898-b791-70d9c015cdb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.342138] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.546241] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52491751-c7cd-b462-5538-f68811def181, 'name': SearchDatastore_Task, 'duration_secs': 0.029064} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.546579] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.546846] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.547096] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.547241] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.547418] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.547688] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e3d4abd-1dde-44a1-b44c-9f2d53639e62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.558089] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.558301] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.558966] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bda17549-c821-4a70-8326-2a7081ffe9b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.565828] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 890.565828] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520e5a27-dd23-9d4a-228b-b53bc907ba8e" [ 890.565828] env[65121]: _type = "Task" [ 890.565828] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.575985] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520e5a27-dd23-9d4a-228b-b53bc907ba8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.612052] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106673, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.621083] env[65121]: DEBUG oslo_vmware.api [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106674, 'name': PowerOffVM_Task, 'duration_secs': 0.333873} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.621382] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.621547] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 890.621879] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0ea0f10-c7e1-44f0-aab6-ed4b0b6ec12e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.677024] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106675, 'name': PowerOffVM_Task, 'duration_secs': 0.320509} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.677024] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.681853] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 890.682165] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a34f9153-c2d3-462c-9e56-5ea469f3dae2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.702374] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 890.702374] env[65121]: value = "task-5106677" [ 890.702374] env[65121]: _type = "Task" [ 890.702374] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.712151] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106677, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.741024] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 890.741279] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 890.741476] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleting the datastore file [datastore2] 93721179-407a-43d9-b0bf-157433a09519 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.741786] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b66df6b7-d404-487c-87a3-1d2b703633fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.752642] env[65121]: DEBUG oslo_vmware.api [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 890.752642] env[65121]: value = "task-5106678" [ 890.752642] env[65121]: _type = "Task" [ 890.752642] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.764376] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 890.768790] env[65121]: DEBUG oslo_vmware.api [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.828374] env[65121]: DEBUG nova.compute.manager [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Received event network-changed-748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 890.828777] env[65121]: DEBUG nova.compute.manager [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Refreshing instance network info cache due to event network-changed-748bfc22-93f1-459b-9b59-18583587dd17. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 890.828982] env[65121]: DEBUG oslo_concurrency.lockutils [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Acquiring lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.829369] env[65121]: DEBUG oslo_concurrency.lockutils [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Acquired lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.829524] env[65121]: DEBUG nova.network.neutron [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Refreshing network info cache for port 748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 890.844692] env[65121]: DEBUG nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 891.077114] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520e5a27-dd23-9d4a-228b-b53bc907ba8e, 'name': SearchDatastore_Task, 'duration_secs': 0.024124} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.077975] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b419dcc4-2836-4fe3-9cae-ef41dd2bbd6f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.084408] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 891.084408] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5233405d-86c0-497b-5b4d-dfdbe94a631a" [ 891.084408] env[65121]: _type = "Task" [ 891.084408] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.093435] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5233405d-86c0-497b-5b4d-dfdbe94a631a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.111351] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106673, 'name': CreateVM_Task, 'duration_secs': 0.595453} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.111577] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.112141] env[65121]: WARNING neutronclient.v2_0.client [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.112550] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.112712] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.113137] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 891.113435] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0f74fd1-50c4-4ba5-bbc7-757afe47bbee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.120025] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 891.120025] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab8d6c-fc40-5419-c713-d20ec363d7dd" [ 891.120025] env[65121]: _type = "Task" [ 891.120025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.130597] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ab8d6c-fc40-5419-c713-d20ec363d7dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.211865] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106677, 'name': ReconfigVM_Task, 'duration_secs': 0.281376} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.212195] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 891.212376] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.212629] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7dcbc346-1993-43ba-9a90-6e4a252302cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.219776] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 891.219776] env[65121]: value = "task-5106679" [ 891.219776] env[65121]: _type = "Task" [ 891.219776] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.228721] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.263312] env[65121]: DEBUG oslo_vmware.api [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177362} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.263585] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.263832] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 891.264106] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 891.264303] env[65121]: INFO nova.compute.manager [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Took 1.19 seconds to destroy the instance on the hypervisor. [ 891.264649] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 891.264849] env[65121]: DEBUG nova.compute.manager [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 891.264943] env[65121]: DEBUG nova.network.neutron [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 891.265199] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.265832] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.266102] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.273509] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 891.273751] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.272s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.273988] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 45.902s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.275926] env[65121]: INFO nova.compute.claims [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.278738] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 891.278895] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Cleaning up deleted instances {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 891.332787] env[65121]: WARNING neutronclient.v2_0.client [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.334464] env[65121]: WARNING openstack [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.335488] env[65121]: WARNING openstack [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.357402] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.368756] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.513054] env[65121]: WARNING openstack [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.513535] env[65121]: WARNING openstack [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.584918] env[65121]: WARNING neutronclient.v2_0.client [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.585704] env[65121]: WARNING openstack [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.586031] env[65121]: WARNING openstack [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.603584] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5233405d-86c0-497b-5b4d-dfdbe94a631a, 'name': SearchDatastore_Task, 'duration_secs': 0.027149} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.603875] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.604153] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.604420] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f67411d7-69ab-4784-818b-664715aebd0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.612423] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 891.612423] env[65121]: value = "task-5106680" [ 891.612423] env[65121]: _type = "Task" [ 891.612423] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.623911] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106680, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.635098] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ab8d6c-fc40-5419-c713-d20ec363d7dd, 'name': SearchDatastore_Task, 'duration_secs': 0.013983} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.640030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.640249] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.640425] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.640567] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.640736] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.641535] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d568e09a-e0c4-4698-a70f-a84395ef7efd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.651972] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.651972] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.652408] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42975c1d-c56a-42a2-beed-1f64cf4387e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.658477] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 891.658477] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523ad5f5-5eca-00e5-667d-dc5ae9c2a07d" [ 891.658477] env[65121]: _type = "Task" [ 891.658477] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.668665] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523ad5f5-5eca-00e5-667d-dc5ae9c2a07d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.685692] env[65121]: DEBUG nova.network.neutron [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Updated VIF entry in instance network info cache for port 748bfc22-93f1-459b-9b59-18583587dd17. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 891.686176] env[65121]: DEBUG nova.network.neutron [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Updating instance_info_cache with network_info: [{"id": "748bfc22-93f1-459b-9b59-18583587dd17", "address": "fa:16:3e:f4:04:0c", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap748bfc22-93", "ovs_interfaceid": "748bfc22-93f1-459b-9b59-18583587dd17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 891.730220] env[65121]: DEBUG oslo_vmware.api [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106679, 'name': PowerOnVM_Task, 'duration_secs': 0.49876} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.730534] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.730775] env[65121]: DEBUG nova.compute.manager [None req-c7cdbbf3-9aa8-4619-9dbd-111fba116aa7 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 891.731902] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e451a69c-c21f-44b7-981b-9bfb2e456b83 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.793511] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] There are 40 instances to clean {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 891.793760] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 75bf0b85-db28-4efa-a517-d7256873b09d] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 892.118678] env[65121]: DEBUG nova.network.neutron [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 892.123734] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106680, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470944} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.124316] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.124551] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.125183] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ecde17e-05bf-489b-a3bc-db710ea2e81b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.133310] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 892.133310] env[65121]: value = "task-5106681" [ 892.133310] env[65121]: _type = "Task" [ 892.133310] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.145808] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106681, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.170085] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523ad5f5-5eca-00e5-667d-dc5ae9c2a07d, 'name': SearchDatastore_Task, 'duration_secs': 0.01286} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.171163] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19a218d6-6d30-4515-80a2-8a08e86735f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.178068] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 892.178068] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cbe149-1ac8-1357-74be-eadb0b881c5d" [ 892.178068] env[65121]: _type = "Task" [ 892.178068] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.186793] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cbe149-1ac8-1357-74be-eadb0b881c5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.189480] env[65121]: DEBUG oslo_concurrency.lockutils [req-19fd84e6-47e5-46c8-ad75-aa052a66d77b req-11f00ed3-703f-40b6-86b3-b3f80422c0c5 service nova] Releasing lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.297371] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: ec8bb91b-69ed-4ee1-ad06-d68bd993c8e7] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 892.624875] env[65121]: INFO nova.compute.manager [-] [instance: 93721179-407a-43d9-b0bf-157433a09519] Took 1.36 seconds to deallocate network for instance. [ 892.643735] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106681, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071544} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.646449] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.647539] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3493339a-e807-4edf-9be0-e008a6ea6072 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.672956] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.677045] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1bbbea9-4fcd-4a5a-b645-05872f063b73 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.703107] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cbe149-1ac8-1357-74be-eadb0b881c5d, 'name': SearchDatastore_Task, 'duration_secs': 0.010635} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.707425] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.707425] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.707782] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 892.707782] env[65121]: value = "task-5106682" [ 892.707782] env[65121]: _type = "Task" [ 892.707782] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.708187] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ea0de50-cb6a-4623-888d-f0766edd2a9b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.719760] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106682, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.724139] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 892.724139] env[65121]: value = "task-5106683" [ 892.724139] env[65121]: _type = "Task" [ 892.724139] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.733810] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106683, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.801717] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 90cfd88c-286e-4c87-bf27-909873359997] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 892.810822] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 892.811129] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 892.811355] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 892.811532] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 892.811741] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.814166] env[65121]: INFO nova.compute.manager [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Terminating instance [ 892.830790] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cac07e0-f2b6-488c-bd7a-4152af79b9b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.842133] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaae85c4-1a75-411f-9034-ded9631e7282 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.875239] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ed54e5-99f8-4db8-9b9b-11e0d877e94d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.878919] env[65121]: DEBUG nova.compute.manager [req-9f51ab66-0123-48bb-b06c-cca54b52f32e req-3fd0bdc8-02ea-4116-9812-2cb1672fe52a service nova] [instance: 93721179-407a-43d9-b0bf-157433a09519] Received event network-vif-deleted-e7651b37-754f-427b-b5a8-5dc1d76d759c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 892.885849] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be961a01-124c-48c8-a3cb-c46ddebd0cdb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.903548] env[65121]: DEBUG nova.compute.provider_tree [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.134260] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 893.220621] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106682, 'name': ReconfigVM_Task, 'duration_secs': 0.303372} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.220969] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866/27940143-16b5-4263-b23c-354ed8ea8866.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.221624] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2c01349-77fa-4a40-bc7c-767b9ce25511 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.231554] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 893.231554] env[65121]: value = "task-5106684" [ 893.231554] env[65121]: _type = "Task" [ 893.231554] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.242851] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106683, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511817} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.243816] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.244204] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.244560] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1fd2523-c56a-4f51-80a7-a72b4fb5b908 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.253061] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106684, 'name': Rename_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.259073] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 893.259073] env[65121]: value = "task-5106685" [ 893.259073] env[65121]: _type = "Task" [ 893.259073] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.271882] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106685, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.304987] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: f694fcb6-053b-4649-ac63-7fa98b1373eb] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 893.321699] env[65121]: DEBUG nova.compute.manager [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 893.321983] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 893.323190] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673f8377-eb2f-45ea-a91a-26ea505f8543 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.332509] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 893.333469] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37210e1c-167e-42a2-961e-9950a9f8ee97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.341369] env[65121]: DEBUG oslo_vmware.api [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 893.341369] env[65121]: value = "task-5106686" [ 893.341369] env[65121]: _type = "Task" [ 893.341369] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.351071] env[65121]: DEBUG oslo_vmware.api [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106686, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.406399] env[65121]: DEBUG nova.scheduler.client.report [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.743338] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106684, 'name': Rename_Task, 'duration_secs': 0.154871} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.743824] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.744062] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dbe5f5f-564a-4808-a2b2-04a819293d64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.751413] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 893.751413] env[65121]: value = "task-5106687" [ 893.751413] env[65121]: _type = "Task" [ 893.751413] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.759771] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106687, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.768460] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106685, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079053} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.768746] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.769568] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00a9fbd-9fec-4fe9-9a4f-0808fef4fb88 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.796178] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.796578] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a544a63-5283-4e91-952a-32b2ab5aceb2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.814036] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: c1632a96-f334-4978-9f31-97f9329577e5] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 893.821054] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 893.821054] env[65121]: value = "task-5106688" [ 893.821054] env[65121]: _type = "Task" [ 893.821054] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.832188] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.854527] env[65121]: DEBUG oslo_vmware.api [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106686, 'name': PowerOffVM_Task, 'duration_secs': 0.1911} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.855648] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.856088] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 893.856425] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-33987118-4d04-4f9f-a862-88bd310c8ba3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.912543] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.638s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.913338] env[65121]: DEBUG nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 893.916848] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 46.552s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 893.935879] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 893.936631] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 893.937089] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Deleting the datastore file [datastore1] 8a23da27-ce6c-4453-9036-65eeeda3ce0d {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.938332] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-373d24e7-b882-466c-b9c9-796f6036b71b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.948881] env[65121]: DEBUG oslo_vmware.api [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 893.948881] env[65121]: value = "task-5106690" [ 893.948881] env[65121]: _type = "Task" [ 893.948881] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.959169] env[65121]: DEBUG oslo_vmware.api [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.262037] env[65121]: DEBUG oslo_vmware.api [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106687, 'name': PowerOnVM_Task, 'duration_secs': 0.490438} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.262363] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.262571] env[65121]: DEBUG nova.compute.manager [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 894.263462] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df0ca86-3775-44e5-af36-a3731bf997eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.316353] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 00362477-c89e-4f60-98a4-d4928081d55e] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 894.332160] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106688, 'name': ReconfigVM_Task, 'duration_secs': 0.306395} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.332523] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.333263] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecac23b6-caa6-4df5-b4b2-d03a5020cae3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.340110] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 894.340110] env[65121]: value = "task-5106691" [ 894.340110] env[65121]: _type = "Task" [ 894.340110] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.350659] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106691, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.420809] env[65121]: DEBUG nova.objects.instance [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lazy-loading 'migration_context' on Instance uuid 293f93f2-c01d-42c8-b1a7-3056805c77de {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.423773] env[65121]: DEBUG nova.compute.utils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 894.425298] env[65121]: DEBUG nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 894.460116] env[65121]: DEBUG oslo_vmware.api [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190159} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.460427] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 894.460559] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 894.460737] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 894.460873] env[65121]: INFO nova.compute.manager [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 894.461143] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 894.461419] env[65121]: DEBUG nova.compute.manager [-] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 894.461419] env[65121]: DEBUG nova.network.neutron [-] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 894.462075] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.462276] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.462814] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.504902] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.782106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.815183] env[65121]: DEBUG nova.compute.manager [req-7cb05255-ad15-4506-84e6-3ac1c4bd138b req-60298b42-0448-4e50-9e7f-3c6f34908c60 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Received event network-vif-deleted-33fba103-6efa-42f4-9689-339edc299410 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 894.815386] env[65121]: INFO nova.compute.manager [req-7cb05255-ad15-4506-84e6-3ac1c4bd138b req-60298b42-0448-4e50-9e7f-3c6f34908c60 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Neutron deleted interface 33fba103-6efa-42f4-9689-339edc299410; detaching it from the instance and deleting it from the info cache [ 894.815577] env[65121]: DEBUG nova.network.neutron [req-7cb05255-ad15-4506-84e6-3ac1c4bd138b req-60298b42-0448-4e50-9e7f-3c6f34908c60 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 894.819585] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 75114f97-fe50-4624-9333-303e411529ea] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 894.854888] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106691, 'name': Rename_Task, 'duration_secs': 0.165021} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.855146] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.855404] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-215ed299-929f-4adf-b075-2b093f1caf1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.862572] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 894.862572] env[65121]: value = "task-5106692" [ 894.862572] env[65121]: _type = "Task" [ 894.862572] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.872587] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.929310] env[65121]: DEBUG nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 895.213019] env[65121]: DEBUG nova.network.neutron [-] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 895.318634] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9cc76b4-81c7-433f-9466-d88ea4ace9c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.322223] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: aab9317b-4ee6-48b3-905b-859a5996f33d] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 895.330798] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdc3b7f-9c26-413b-8a4b-78a46537c468 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.379933] env[65121]: DEBUG nova.compute.manager [req-7cb05255-ad15-4506-84e6-3ac1c4bd138b req-60298b42-0448-4e50-9e7f-3c6f34908c60 service nova] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Detach interface failed, port_id=33fba103-6efa-42f4-9689-339edc299410, reason: Instance 8a23da27-ce6c-4453-9036-65eeeda3ce0d could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 895.395294] env[65121]: DEBUG oslo_vmware.api [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106692, 'name': PowerOnVM_Task, 'duration_secs': 0.502923} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.395876] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.396321] env[65121]: INFO nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Took 8.85 seconds to spawn the instance on the hypervisor. [ 895.396827] env[65121]: DEBUG nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 895.398441] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54675d17-7863-46d1-bdb9-ee7b6fcc4203 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.514115] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228e77bb-61cb-43d0-9c02-17dc7cc3bc44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.524679] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec50577-12e1-46a1-ae66-26338b68b17f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.559397] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2648cb18-9c1f-40ee-8b90-31190031b91a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.567750] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f668eacc-95bd-4c21-a3c9-877c8f7adcad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.583119] env[65121]: DEBUG nova.compute.provider_tree [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.715996] env[65121]: INFO nova.compute.manager [-] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Took 1.25 seconds to deallocate network for instance. [ 895.826134] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 19b1f3f9-842e-4150-8890-b0b22393c3af] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 895.927871] env[65121]: INFO nova.compute.manager [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Took 53.95 seconds to build instance. [ 895.942070] env[65121]: DEBUG nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 895.944482] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.944704] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.945115] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.945192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.945486] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 895.947293] env[65121]: INFO nova.compute.manager [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Terminating instance [ 895.972770] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 895.973054] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 895.973205] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 895.974317] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 895.974317] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 895.974317] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 895.974317] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.974317] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 895.974317] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 895.974317] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 895.974550] env[65121]: DEBUG nova.virt.hardware [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 895.975672] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291fd88e-0f97-4104-acad-1c6a23760cc1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.984810] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80571676-4911-44a2-b72b-decb0ab51d15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.002390] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.008380] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 896.008669] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.008883] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2efb2d38-b6f9-4637-a715-8e82aacf4b60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.027957] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.027957] env[65121]: value = "task-5106693" [ 896.027957] env[65121]: _type = "Task" [ 896.027957] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.039036] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106693, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.089605] env[65121]: DEBUG nova.scheduler.client.report [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 896.224777] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 896.317486] env[65121]: DEBUG nova.compute.manager [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Received event network-changed-748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 896.317721] env[65121]: DEBUG nova.compute.manager [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Refreshing instance network info cache due to event network-changed-748bfc22-93f1-459b-9b59-18583587dd17. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 896.317910] env[65121]: DEBUG oslo_concurrency.lockutils [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Acquiring lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.318057] env[65121]: DEBUG oslo_concurrency.lockutils [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Acquired lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 896.318237] env[65121]: DEBUG nova.network.neutron [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Refreshing network info cache for port 748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 896.330810] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 41af3ee5-0ee6-4957-bad0-5aef51d55bd5] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 896.430082] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43e6948e-3b7c-4a02-b294-726f836e1aaa tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.347s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.453702] env[65121]: DEBUG nova.compute.manager [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 896.453948] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.454962] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-005792c7-9f09-44f8-9871-684bb6659464 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.463985] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.464291] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10487ec8-90c1-4296-862f-e35da0ccebe8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.472930] env[65121]: DEBUG oslo_vmware.api [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 896.472930] env[65121]: value = "task-5106694" [ 896.472930] env[65121]: _type = "Task" [ 896.472930] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.483761] env[65121]: DEBUG oslo_vmware.api [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.540216] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106693, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.824375] env[65121]: WARNING neutronclient.v2_0.client [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 896.824375] env[65121]: WARNING openstack [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.824375] env[65121]: WARNING openstack [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.833906] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 3f69040e-7df3-4535-a3a7-90f3348ef346] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 896.966970] env[65121]: WARNING openstack [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 896.967379] env[65121]: WARNING openstack [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 896.992622] env[65121]: DEBUG oslo_vmware.api [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106694, 'name': PowerOffVM_Task, 'duration_secs': 0.281591} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.992622] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 896.992622] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 896.992622] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12fdb2f9-dc84-4761-9f92-06c7ff3bf9aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.042055] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106693, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.044174] env[65121]: WARNING neutronclient.v2_0.client [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.044884] env[65121]: WARNING openstack [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.045911] env[65121]: WARNING openstack [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.075290] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.075611] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.075797] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleting the datastore file [datastore1] 99b1ff9c-38b7-4947-81bb-d200a3b2a09a {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.076114] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57121978-0d56-4dab-95b2-01b865eb5cd5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.083543] env[65121]: DEBUG oslo_vmware.api [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 897.083543] env[65121]: value = "task-5106696" [ 897.083543] env[65121]: _type = "Task" [ 897.083543] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.092970] env[65121]: DEBUG oslo_vmware.api [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106696, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.105325] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.188s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.111555] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 45.227s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 897.111775] env[65121]: DEBUG nova.objects.instance [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 897.148450] env[65121]: DEBUG nova.network.neutron [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Updated VIF entry in instance network info cache for port 748bfc22-93f1-459b-9b59-18583587dd17. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 897.149170] env[65121]: DEBUG nova.network.neutron [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Updating instance_info_cache with network_info: [{"id": "748bfc22-93f1-459b-9b59-18583587dd17", "address": "fa:16:3e:f4:04:0c", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap748bfc22-93", "ovs_interfaceid": "748bfc22-93f1-459b-9b59-18583587dd17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 897.338199] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 50083ee5-9655-4cab-9d50-04a97baac626] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 897.541341] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106693, 'name': CreateVM_Task, 'duration_secs': 1.319637} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.541570] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.541977] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.542149] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 897.542473] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 897.542839] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50bdb19d-204e-495d-b034-f01d098806cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.548503] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 897.548503] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52baa2c3-5426-bf69-6902-95e19517039b" [ 897.548503] env[65121]: _type = "Task" [ 897.548503] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.558095] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52baa2c3-5426-bf69-6902-95e19517039b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.593528] env[65121]: DEBUG oslo_vmware.api [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106696, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.293152} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.593871] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.594137] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.594317] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.594502] env[65121]: INFO nova.compute.manager [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 897.594763] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 897.594958] env[65121]: DEBUG nova.compute.manager [-] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 897.595081] env[65121]: DEBUG nova.network.neutron [-] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 897.595336] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.595901] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.596176] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.646958] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.651763] env[65121]: DEBUG oslo_concurrency.lockutils [req-39a079c2-d514-4d8e-b6ac-09b0092d794c req-531a0b43-fc53-4156-9c40-cdff859baf31 service nova] Releasing lock "refresh_cache-274d0ccd-c707-4a68-b280-16de2bc74d73" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 897.843000] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 5b9c3d1a-c314-4a12-8fe3-ce8ed596fb5f] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 897.909103] env[65121]: DEBUG nova.compute.manager [req-89e07c05-47f7-46cb-86b4-2055110609d4 req-8355012d-6907-496f-acc0-78a9a59ed305 service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Received event network-vif-deleted-027b9031-f921-43b2-8580-de8d028869c3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 897.909575] env[65121]: INFO nova.compute.manager [req-89e07c05-47f7-46cb-86b4-2055110609d4 req-8355012d-6907-496f-acc0-78a9a59ed305 service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Neutron deleted interface 027b9031-f921-43b2-8580-de8d028869c3; detaching it from the instance and deleting it from the info cache [ 897.909575] env[65121]: DEBUG nova.network.neutron [req-89e07c05-47f7-46cb-86b4-2055110609d4 req-8355012d-6907-496f-acc0-78a9a59ed305 service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.060399] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52baa2c3-5426-bf69-6902-95e19517039b, 'name': SearchDatastore_Task, 'duration_secs': 0.035319} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.060754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.060985] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.061235] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.061374] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 898.061566] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.061833] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a30f7d5c-3892-4a50-885c-20b8e8250add {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.072083] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.072083] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.072505] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25dd3ade-21d2-4261-ad87-a24c8ac7fab7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.078599] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 898.078599] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525f8ed4-afaf-da73-bb03-68193a10b825" [ 898.078599] env[65121]: _type = "Task" [ 898.078599] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.088189] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525f8ed4-afaf-da73-bb03-68193a10b825, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.127729] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74eb1742-f18b-492e-ace2-b54d752b9470 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.131382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 44.315s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.131567] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.133852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.209s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.135937] env[65121]: INFO nova.compute.claims [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.171032] env[65121]: INFO nova.scheduler.client.report [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Deleted allocations for instance f1a5fdce-a940-43d4-83d7-716786c9ac34 [ 898.346148] env[65121]: DEBUG nova.network.neutron [-] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.348128] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 06a99a15-5c8a-4be0-b393-80a104b9ba76] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 898.412682] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7c2ae36-8dd7-4a14-86c5-842b60637711 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.423429] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a7ebb2-d029-45f3-84c8-a6d4aff896ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.462697] env[65121]: DEBUG nova.compute.manager [req-89e07c05-47f7-46cb-86b4-2055110609d4 req-8355012d-6907-496f-acc0-78a9a59ed305 service nova] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Detach interface failed, port_id=027b9031-f921-43b2-8580-de8d028869c3, reason: Instance 99b1ff9c-38b7-4947-81bb-d200a3b2a09a could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 898.589117] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525f8ed4-afaf-da73-bb03-68193a10b825, 'name': SearchDatastore_Task, 'duration_secs': 0.01339} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.589911] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a00c15cc-b742-4872-9121-4997effad7d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.596081] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 898.596081] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521b06af-85ef-e5da-11e0-58283e6a88ea" [ 898.596081] env[65121]: _type = "Task" [ 898.596081] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.604655] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521b06af-85ef-e5da-11e0-58283e6a88ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.660860] env[65121]: INFO nova.compute.manager [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Swapping old allocation on dict_keys(['d56783bf-3ede-475a-8c5a-8d8303049e47']) held by migration 2ce31b97-0aef-4465-995a-13b22493087b for instance [ 898.681574] env[65121]: DEBUG oslo_concurrency.lockutils [None req-308292a8-64fa-4203-a9ba-19e413c174ee tempest-SecurityGroupsTestJSON-1711894211 tempest-SecurityGroupsTestJSON-1711894211-project-member] Lock "f1a5fdce-a940-43d4-83d7-716786c9ac34" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.044s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.691990] env[65121]: DEBUG nova.scheduler.client.report [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Overwriting current allocation {'allocations': {'d56783bf-3ede-475a-8c5a-8d8303049e47': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 97}}, 'project_id': '982f66705583488cb369f19160cc2ee5', 'user_id': '0f1ca61e513f4855a105e40ff37a2d75', 'consumer_generation': 1} on consumer 293f93f2-c01d-42c8-b1a7-3056805c77de {{(pid=65121) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 898.747056] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 898.797232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.797474] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquired lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 898.797592] env[65121]: DEBUG nova.network.neutron [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 898.848728] env[65121]: INFO nova.compute.manager [-] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Took 1.25 seconds to deallocate network for instance. [ 898.850848] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: c1004635-b318-489c-9e16-6cb545279953] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 899.107450] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521b06af-85ef-e5da-11e0-58283e6a88ea, 'name': SearchDatastore_Task, 'duration_secs': 0.017871} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.107761] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 899.108030] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.108298] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3c061e4-dfbf-4ec1-870a-705d36eb71ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.115308] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 899.115308] env[65121]: value = "task-5106697" [ 899.115308] env[65121]: _type = "Task" [ 899.115308] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.126193] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106697, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.302385] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.302385] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.302385] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.355030] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: ab27fa56-f672-4096-a8f4-cce5ff4d5460] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 899.358030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.448317] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.448744] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.515902] env[65121]: WARNING neutronclient.v2_0.client [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.517669] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.519107] env[65121]: WARNING openstack [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.633823] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106697, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.656748] env[65121]: DEBUG nova.network.neutron [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance_info_cache with network_info: [{"id": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "address": "fa:16:3e:c8:d8:39", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79bfb570-6e", "ovs_interfaceid": "79bfb570-6ed2-4c0f-83e1-1f2cfb088e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 899.732017] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1400442a-356f-46ff-b233-5a2187e4205d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.743112] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83226d30-5063-4941-855a-dbdd2604f3be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.783058] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c75bfe-57ce-46cf-ad76-bed0d7d5b8b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.796022] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a251e3-1878-472b-9268-18bd9bc4206d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.809054] env[65121]: DEBUG nova.compute.provider_tree [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.858977] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 6164d64c-75c5-469e-ac32-667842b3717a] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 900.133992] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106697, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.800237} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.133992] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 900.133992] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.133992] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d31fffe-1b78-4227-a0a8-d277223ea190 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.141632] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 900.141632] env[65121]: value = "task-5106698" [ 900.141632] env[65121]: _type = "Task" [ 900.141632] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.151839] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106698, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.165796] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Releasing lock "refresh_cache-293f93f2-c01d-42c8-b1a7-3056805c77de" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 900.166249] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.166581] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99522fd5-da72-4fc6-a3d7-7b00d493823d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.178533] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 900.178533] env[65121]: value = "task-5106699" [ 900.178533] env[65121]: _type = "Task" [ 900.178533] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.190882] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.313053] env[65121]: DEBUG nova.scheduler.client.report [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 900.363246] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: b1ad9d15-7ef4-404a-9751-ecdd02ea768b] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 900.652127] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106698, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073336} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.652426] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.653256] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2343c938-5d3d-4fda-8b49-950c8ecbdc6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.675591] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.675921] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4725aec8-8eeb-44ed-82f5-d002dcf29ef6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.700786] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106699, 'name': PowerOffVM_Task, 'duration_secs': 0.464546} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.702350] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.703060] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:22:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='8a66939b-83da-49a6-8c3f-9ca9e786ade0',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1368792546',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 900.703280] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 900.703437] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 900.703778] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 900.703969] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 900.704142] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 900.704354] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.704512] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 900.704917] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 900.704917] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 900.705031] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 900.710652] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 900.710652] env[65121]: value = "task-5106700" [ 900.710652] env[65121]: _type = "Task" [ 900.710652] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.710912] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4502215-d07a-40fa-af23-bfb9ac9a1d47 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.735048] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 900.735048] env[65121]: value = "task-5106701" [ 900.735048] env[65121]: _type = "Task" [ 900.735048] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.735303] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106700, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.744953] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106701, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.820177] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.821048] env[65121]: DEBUG nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 900.824258] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.573s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.824258] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.826559] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 40.293s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.826795] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.828945] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.584s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.830390] env[65121]: INFO nova.compute.claims [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.868688] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 1e4a9d58-f848-4ff2-8558-4f23c6082a5f] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 900.895463] env[65121]: INFO nova.scheduler.client.report [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Deleted allocations for instance e2e03e5c-4a71-4555-9cd4-d67e178064ec [ 900.899778] env[65121]: INFO nova.scheduler.client.report [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Deleted allocations for instance c993d5a4-49d6-43aa-a1f0-4aac91fad953 [ 901.233584] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106700, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.247116] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106701, 'name': ReconfigVM_Task, 'duration_secs': 0.30399} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.248013] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753b6e09-7398-4fd0-b76b-0d56b9df63a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.271130] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:22:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='8a66939b-83da-49a6-8c3f-9ca9e786ade0',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1368792546',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 901.272139] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 901.272240] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 901.272400] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 901.272635] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 901.272843] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 901.273722] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.273722] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 901.273889] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 901.273924] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 901.274504] env[65121]: DEBUG nova.virt.hardware [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 901.274969] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47c9e366-bd9e-4c8d-8549-4265ec487353 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.284612] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 901.284612] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d8cb6f-4266-8e59-69c1-c62c5d0a79d5" [ 901.284612] env[65121]: _type = "Task" [ 901.284612] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.295852] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d8cb6f-4266-8e59-69c1-c62c5d0a79d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.337200] env[65121]: DEBUG nova.compute.utils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 901.338747] env[65121]: DEBUG nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 901.339582] env[65121]: DEBUG nova.network.neutron [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 901.339964] env[65121]: WARNING neutronclient.v2_0.client [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.339964] env[65121]: WARNING neutronclient.v2_0.client [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.340553] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.341032] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.377493] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: cd86f815-4b46-45b1-9a1c-ab2e28f63626] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 901.410199] env[65121]: DEBUG nova.policy [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26beb81d28f44cc8932ea8e87a4cd2c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0630960dcbf44781be05184565d81932', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 901.416551] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18987c1e-01f4-48aa-9292-ebd3d5562f79 tempest-ServerShowV257Test-527026045 tempest-ServerShowV257Test-527026045-project-member] Lock "e2e03e5c-4a71-4555-9cd4-d67e178064ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.440s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.419292] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d511ce70-4b5c-4891-928b-4fbcb94f980d tempest-ListImageFiltersTestJSON-547404429 tempest-ListImageFiltersTestJSON-547404429-project-member] Lock "c993d5a4-49d6-43aa-a1f0-4aac91fad953" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.384s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.703117] env[65121]: DEBUG nova.network.neutron [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Successfully created port: 9e8f421a-1606-4c5a-821d-90bee69ab953 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 901.734467] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106700, 'name': ReconfigVM_Task, 'duration_secs': 0.549971} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.734467] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.734798] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4167a8f2-a11a-4f83-95e8-e0def7d6e846 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.741828] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 901.741828] env[65121]: value = "task-5106702" [ 901.741828] env[65121]: _type = "Task" [ 901.741828] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.752121] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106702, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.796591] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d8cb6f-4266-8e59-69c1-c62c5d0a79d5, 'name': SearchDatastore_Task, 'duration_secs': 0.012776} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.803494] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfiguring VM instance instance-0000002a to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 901.804117] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ab84065-81f2-4e93-9aac-cccc97ac9502 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.827738] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 901.827738] env[65121]: value = "task-5106703" [ 901.827738] env[65121]: _type = "Task" [ 901.827738] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.837932] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106703, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.851537] env[65121]: DEBUG nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 901.883562] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 124876fe-a4f2-4140-8dee-d8bb72b5c9a5] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 902.253971] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106702, 'name': Rename_Task, 'duration_secs': 0.191084} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.254277] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.254575] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f6d6fdb-18e2-4cc8-b0c3-0aed79b80718 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.261230] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 902.261230] env[65121]: value = "task-5106704" [ 902.261230] env[65121]: _type = "Task" [ 902.261230] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.272349] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.341939] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106703, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.386617] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d77ac4-1a81-486e-9ea3-97e837d21bff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.394822] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 4b6fe349-335e-4202-ab76-04f6ccc036ee] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 902.397598] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5bf865-e5cc-4008-a7b1-3821a5b0ab52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.432251] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a02fd17-10d4-44c4-a275-853211c00e27 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.441022] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b339d9b5-6e48-4556-8a9c-ac64bd5037d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.455344] env[65121]: DEBUG nova.compute.provider_tree [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.775489] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106704, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.841321] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106703, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.862793] env[65121]: DEBUG nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 902.889580] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 902.890016] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 902.890559] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 902.890559] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 902.890635] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 902.890770] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 902.890990] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.891158] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 902.891321] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 902.891474] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 902.891640] env[65121]: DEBUG nova.virt.hardware [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 902.892553] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a43cc8f-bd7d-4217-bf86-c814bc0e199d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.902534] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad63027-c854-4e25-bc1f-b4010536d420 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.907480] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 61bd5995-701e-430f-9aae-4b266089e313] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 902.958714] env[65121]: DEBUG nova.scheduler.client.report [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.275442] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106704, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.342482] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106703, 'name': ReconfigVM_Task, 'duration_secs': 1.218769} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.342965] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfigured VM instance instance-0000002a to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 903.343869] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb83ec7a-c85a-4342-8440-22f77c7369ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.369774] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.370545] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6159febc-be65-4680-b368-61d0d0a1003e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.391837] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 903.391837] env[65121]: value = "task-5106705" [ 903.391837] env[65121]: _type = "Task" [ 903.391837] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.402938] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.410813] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: b9c78b3b-dd43-4199-98a7-ff57fd129f43] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 903.421689] env[65121]: DEBUG nova.network.neutron [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Successfully updated port: 9e8f421a-1606-4c5a-821d-90bee69ab953 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 903.464524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.465063] env[65121]: DEBUG nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 903.468414] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.890s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.470423] env[65121]: INFO nova.compute.claims [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.572501] env[65121]: DEBUG nova.compute.manager [req-438a2cf3-ad47-4977-bc19-2de76156f23d req-b8ef0209-3029-4a44-bc70-6a2607875c6c service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Received event network-vif-plugged-9e8f421a-1606-4c5a-821d-90bee69ab953 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 903.572765] env[65121]: DEBUG oslo_concurrency.lockutils [req-438a2cf3-ad47-4977-bc19-2de76156f23d req-b8ef0209-3029-4a44-bc70-6a2607875c6c service nova] Acquiring lock "df5abcb5-583f-4b28-a074-3a3221d74d87-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.572969] env[65121]: DEBUG oslo_concurrency.lockutils [req-438a2cf3-ad47-4977-bc19-2de76156f23d req-b8ef0209-3029-4a44-bc70-6a2607875c6c service nova] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.573332] env[65121]: DEBUG oslo_concurrency.lockutils [req-438a2cf3-ad47-4977-bc19-2de76156f23d req-b8ef0209-3029-4a44-bc70-6a2607875c6c service nova] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.573568] env[65121]: DEBUG nova.compute.manager [req-438a2cf3-ad47-4977-bc19-2de76156f23d req-b8ef0209-3029-4a44-bc70-6a2607875c6c service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] No waiting events found dispatching network-vif-plugged-9e8f421a-1606-4c5a-821d-90bee69ab953 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 903.573766] env[65121]: WARNING nova.compute.manager [req-438a2cf3-ad47-4977-bc19-2de76156f23d req-b8ef0209-3029-4a44-bc70-6a2607875c6c service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Received unexpected event network-vif-plugged-9e8f421a-1606-4c5a-821d-90bee69ab953 for instance with vm_state building and task_state spawning. [ 903.775358] env[65121]: DEBUG oslo_vmware.api [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106704, 'name': PowerOnVM_Task, 'duration_secs': 1.338936} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.775610] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.775820] env[65121]: INFO nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Took 7.83 seconds to spawn the instance on the hypervisor. [ 903.775992] env[65121]: DEBUG nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 903.776781] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752a15e5-7b49-4147-a00e-d1d419a7dad6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.902831] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.914967] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 6218957b-6329-4004-97ca-07231b55937a] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 903.927219] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-df5abcb5-583f-4b28-a074-3a3221d74d87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.927471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-df5abcb5-583f-4b28-a074-3a3221d74d87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 903.927670] env[65121]: DEBUG nova.network.neutron [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 903.975285] env[65121]: DEBUG nova.compute.utils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 903.979015] env[65121]: DEBUG nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 904.296147] env[65121]: INFO nova.compute.manager [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Took 58.95 seconds to build instance. [ 904.405709] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106705, 'name': ReconfigVM_Task, 'duration_secs': 0.61414} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.405967] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de/293f93f2-c01d-42c8-b1a7-3056805c77de.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.406863] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a72361-07f6-47b3-8adc-742d516c31bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.425554] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 7a0d0769-31c5-4ae1-8520-a744f64d39c6] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 904.428576] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbf67f5-2016-4fff-8af0-1bb86e0197c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.448640] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.449017] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.459265] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe7e263-e9b8-4ee0-9587-c3e5bb6b4627 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.483583] env[65121]: DEBUG nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 904.489189] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6790af3-a3d1-434e-91de-e994a9990b2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.495739] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.496036] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e775205-7588-4dbc-9f3d-a6c28c035074 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.504648] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 904.504648] env[65121]: value = "task-5106706" [ 904.504648] env[65121]: _type = "Task" [ 904.504648] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.519104] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.563624] env[65121]: DEBUG nova.network.neutron [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 904.763254] env[65121]: INFO nova.compute.manager [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Rebuilding instance [ 904.773784] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.774375] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.799535] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13fddc28-757a-4370-8fe0-2de2e5649748 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.561s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 904.818657] env[65121]: DEBUG nova.compute.manager [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 904.819818] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fb21e1-3c82-4c6d-914c-a9e5f7add3e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.910198] env[65121]: WARNING neutronclient.v2_0.client [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.911160] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.911815] env[65121]: WARNING openstack [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.931701] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 6a4841b6-6076-4e9c-a6cf-a658957e2931] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 905.020642] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106706, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.124548] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff857c3-f040-4b7d-b540-62cfdab8c52a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.134989] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be4ba07-51ac-4bff-8afe-992dbcb62353 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.172858] env[65121]: DEBUG nova.network.neutron [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Updating instance_info_cache with network_info: [{"id": "9e8f421a-1606-4c5a-821d-90bee69ab953", "address": "fa:16:3e:42:8c:ac", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e8f421a-16", "ovs_interfaceid": "9e8f421a-1606-4c5a-821d-90bee69ab953", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 905.175550] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d66411a-be57-489a-b0cb-75c2bf6dfa04 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.187089] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b06150-2c4a-4946-ab1b-c5ff3f66ab0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.210508] env[65121]: DEBUG nova.compute.provider_tree [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.440439] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: e2d32a5e-c350-4b2b-9243-c3b412193a82] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 905.497276] env[65121]: DEBUG nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 905.517979] env[65121]: DEBUG oslo_vmware.api [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106706, 'name': PowerOnVM_Task, 'duration_secs': 0.750166} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.518487] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.534573] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 905.534877] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.535045] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 905.535234] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.535374] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 905.535562] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 905.535787] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.535942] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 905.536130] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 905.536284] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 905.536451] env[65121]: DEBUG nova.virt.hardware [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 905.537340] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f5ecea-fcca-44de-8d3e-e947cc9d0067 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.547010] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f36d36-08e8-47f1-bef2-19545fb0abe6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.563918] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.569679] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Creating folder: Project (6e8e753331e14e838d3870589d549cac). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.570069] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2dd9248-df16-4b57-9f94-ad5cd20190b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.583183] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Created folder: Project (6e8e753331e14e838d3870589d549cac) in parent group-v993268. [ 905.583376] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Creating folder: Instances. Parent ref: group-v993463. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.583622] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46d47f18-710f-4c5f-a12a-666a6853c105 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.594168] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Created folder: Instances in parent group-v993463. [ 905.595040] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 905.595040] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.595040] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a0f2246-0f8f-4626-8d7e-fbee825d05a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.615399] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.615399] env[65121]: value = "task-5106709" [ 905.615399] env[65121]: _type = "Task" [ 905.615399] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.624459] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106709, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.678028] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-df5abcb5-583f-4b28-a074-3a3221d74d87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.678469] env[65121]: DEBUG nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Instance network_info: |[{"id": "9e8f421a-1606-4c5a-821d-90bee69ab953", "address": "fa:16:3e:42:8c:ac", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e8f421a-16", "ovs_interfaceid": "9e8f421a-1606-4c5a-821d-90bee69ab953", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 905.679356] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:8c:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e8f421a-1606-4c5a-821d-90bee69ab953', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.690777] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 905.692591] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.694052] env[65121]: DEBUG nova.compute.manager [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Received event network-changed-9e8f421a-1606-4c5a-821d-90bee69ab953 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 905.696410] env[65121]: DEBUG nova.compute.manager [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Refreshing instance network info cache due to event network-changed-9e8f421a-1606-4c5a-821d-90bee69ab953. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 905.696410] env[65121]: DEBUG oslo_concurrency.lockutils [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Acquiring lock "refresh_cache-df5abcb5-583f-4b28-a074-3a3221d74d87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.696410] env[65121]: DEBUG oslo_concurrency.lockutils [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Acquired lock "refresh_cache-df5abcb5-583f-4b28-a074-3a3221d74d87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.696410] env[65121]: DEBUG nova.network.neutron [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Refreshing network info cache for port 9e8f421a-1606-4c5a-821d-90bee69ab953 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 905.696588] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5187f48e-2659-4841-b717-a9c334f33eb7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.716470] env[65121]: DEBUG nova.scheduler.client.report [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.731642] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.731642] env[65121]: value = "task-5106710" [ 905.731642] env[65121]: _type = "Task" [ 905.731642] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.741056] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106710, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.848625] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.849396] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5213c7e2-fb3f-4997-a796-060b45866b0d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.859738] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 905.859738] env[65121]: value = "task-5106711" [ 905.859738] env[65121]: _type = "Task" [ 905.859738] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.871726] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106711, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.944112] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 5854d21a-d1a8-4043-aec8-b37ff25c40e7] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 906.134331] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106709, 'name': CreateVM_Task, 'duration_secs': 0.379448} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.134566] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.135068] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.136172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.136745] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 906.137698] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e23710b-6548-4625-9963-6586e7896f80 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.149238] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 906.149238] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52431910-cb5e-1c11-3ab7-add97662f973" [ 906.149238] env[65121]: _type = "Task" [ 906.149238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.158964] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52431910-cb5e-1c11-3ab7-add97662f973, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.217167] env[65121]: WARNING neutronclient.v2_0.client [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.218041] env[65121]: WARNING openstack [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.218518] env[65121]: WARNING openstack [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.233182] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.764s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.233182] env[65121]: DEBUG nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 906.239515] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 40.085s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.239515] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.244789] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 34.963s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.261636] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106710, 'name': CreateVM_Task, 'duration_secs': 0.409178} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.261636] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.262117] env[65121]: WARNING neutronclient.v2_0.client [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.263178] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.289050] env[65121]: INFO nova.scheduler.client.report [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Deleted allocations for instance 5d352184-7074-4fab-b4ac-cc11a6b936b1 [ 906.371646] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106711, 'name': PowerOffVM_Task, 'duration_secs': 0.139783} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.376394] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.376762] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.377645] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0115b92d-5903-4afe-9379-cc564668b872 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.385723] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.385723] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-299c2d87-6285-44c1-9c58-8dfd686476a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.417113] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.420020] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.420020] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleting the datastore file [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.422023] env[65121]: WARNING openstack [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.422388] env[65121]: WARNING openstack [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.429270] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45455f7e-b6d1-4205-8cb2-c4a9348ae832 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.437346] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 906.437346] env[65121]: value = "task-5106713" [ 906.437346] env[65121]: _type = "Task" [ 906.437346] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.447990] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 46f89cfb-14aa-483d-aa8a-bc91356272ca] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 906.450660] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106713, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.534678] env[65121]: INFO nova.compute.manager [None req-cf6fffa0-e12b-46f1-ba12-26d35809e3f1 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance to original state: 'active' [ 906.662351] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52431910-cb5e-1c11-3ab7-add97662f973, 'name': SearchDatastore_Task, 'duration_secs': 0.01137} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.663295] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.663295] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.663295] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.663446] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.663557] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.663772] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.664113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 906.664500] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c503b80d-aa85-4a8a-8f2f-479d1d983eb4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.666542] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01ab7d61-b6f1-41cd-9233-4ff8aa00e35c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.673025] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 906.673025] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524ae485-4344-1664-4cc6-22b85cad76f0" [ 906.673025] env[65121]: _type = "Task" [ 906.673025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.677881] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.678084] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.679508] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcea3b7a-9080-4e3e-863f-8a1a56c4c95d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.686032] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524ae485-4344-1664-4cc6-22b85cad76f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.690312] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 906.690312] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52801ef9-c422-2927-154c-ac9ab5f39234" [ 906.690312] env[65121]: _type = "Task" [ 906.690312] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.703037] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52801ef9-c422-2927-154c-ac9ab5f39234, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.755887] env[65121]: INFO nova.compute.claims [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.761513] env[65121]: DEBUG nova.compute.utils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 906.763427] env[65121]: DEBUG nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 906.763749] env[65121]: DEBUG nova.network.neutron [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 906.764168] env[65121]: WARNING neutronclient.v2_0.client [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.764584] env[65121]: WARNING neutronclient.v2_0.client [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.765354] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.765829] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.796934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-915c09a8-9a3f-4e8f-8bc5-619feedc0128 tempest-ServerMetadataNegativeTestJSON-807233710 tempest-ServerMetadataNegativeTestJSON-807233710-project-member] Lock "5d352184-7074-4fab-b4ac-cc11a6b936b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.150s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.836505] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.836777] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.948687] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106713, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108166} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.949226] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.949226] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 906.949355] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 906.952317] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 4ffc659b-b7a9-42ac-bcd0-0cae8b0ea818] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 907.186388] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524ae485-4344-1664-4cc6-22b85cad76f0, 'name': SearchDatastore_Task, 'duration_secs': 0.020921} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.186734] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.187010] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.187443] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.189969] env[65121]: WARNING neutronclient.v2_0.client [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 907.191175] env[65121]: WARNING openstack [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.191175] env[65121]: WARNING openstack [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.210827] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52801ef9-c422-2927-154c-ac9ab5f39234, 'name': SearchDatastore_Task, 'duration_secs': 0.01207} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.212055] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d80705ea-6ae1-46c3-bb0f-b4cdcfa119d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.220357] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 907.220357] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5216772c-a04f-3e76-d416-d060571aea6c" [ 907.220357] env[65121]: _type = "Task" [ 907.220357] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.230407] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5216772c-a04f-3e76-d416-d060571aea6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.264915] env[65121]: INFO nova.compute.resource_tracker [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating resource usage from migration 71f1858e-40ac-4394-b0fc-2d77e0195ba7 [ 907.277298] env[65121]: DEBUG nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 907.304901] env[65121]: DEBUG nova.policy [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8931a266fddc4346b0c5243608a39bff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00383aa3355e438cb703c2b86c7917f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 907.339094] env[65121]: DEBUG nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 907.459057] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 6887bb66-7186-4d58-be59-d0c53c3221c6] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 907.734987] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5216772c-a04f-3e76-d416-d060571aea6c, 'name': SearchDatastore_Task, 'duration_secs': 0.013144} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.738282] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.738606] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.740502] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 907.740502] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.741484] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-793d0504-b63c-444b-b4d5-239d420eceb6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.743212] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc0d74ed-aa47-4dd6-adcf-d315632d6563 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.754618] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 907.754618] env[65121]: value = "task-5106714" [ 907.754618] env[65121]: _type = "Task" [ 907.754618] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.756509] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.756939] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.764086] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14a097d-7b76-4912-8b08-5a20399b4bf8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.773232] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 907.773232] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520adbf4-f484-90c8-da4b-7f5e250cea31" [ 907.773232] env[65121]: _type = "Task" [ 907.773232] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.777618] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106714, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.796367] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520adbf4-f484-90c8-da4b-7f5e250cea31, 'name': SearchDatastore_Task, 'duration_secs': 0.012238} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.797758] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be4c9f5c-c971-4407-a605-159b15338b1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.806107] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 907.806107] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e3fe14-958b-cebd-cda2-0d38055db7e1" [ 907.806107] env[65121]: _type = "Task" [ 907.806107] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.820205] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e3fe14-958b-cebd-cda2-0d38055db7e1, 'name': SearchDatastore_Task} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.822705] env[65121]: DEBUG nova.network.neutron [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Updated VIF entry in instance network info cache for port 9e8f421a-1606-4c5a-821d-90bee69ab953. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 907.822929] env[65121]: DEBUG nova.network.neutron [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Updating instance_info_cache with network_info: [{"id": "9e8f421a-1606-4c5a-821d-90bee69ab953", "address": "fa:16:3e:42:8c:ac", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e8f421a-16", "ovs_interfaceid": "9e8f421a-1606-4c5a-821d-90bee69ab953", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 907.824175] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.825013] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] df5abcb5-583f-4b28-a074-3a3221d74d87/df5abcb5-583f-4b28-a074-3a3221d74d87.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.827670] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d150fdad-193d-4778-ab0b-414b55e5ad24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.837913] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 907.837913] env[65121]: value = "task-5106715" [ 907.837913] env[65121]: _type = "Task" [ 907.837913] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.857802] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106715, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.869908] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 907.952574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa097d56-3e77-428f-b93d-060495b2912d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.965701] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 5e915c8e-150a-4bfd-8005-a7a1a119057d] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 907.969136] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fc1dce-0327-4844-8ffc-eb2b7aafcbcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.011749] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0c0daf-ea89-4ac1-8a68-58283a0fac8d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.024593] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 908.024906] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 908.025775] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 908.026363] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 908.026363] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 908.026455] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 908.026701] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.026957] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 908.027185] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 908.027397] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 908.027608] env[65121]: DEBUG nova.virt.hardware [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 908.028928] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138a90fd-5b91-481e-b75c-0da401731968 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.038411] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e701dbbb-7175-499b-a202-b09f955f9571 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.048391] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667b5b5c-04f2-4271-8093-5b362ea25ef9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.062459] env[65121]: DEBUG nova.compute.provider_tree [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.077725] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.084349] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 908.085305] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.085366] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e18e052-f356-417b-986a-93fddeefd45a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.102415] env[65121]: DEBUG nova.network.neutron [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Successfully created port: c2c37a16-3107-4510-9c9d-17bf130945ff {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 908.113387] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.113387] env[65121]: value = "task-5106716" [ 908.113387] env[65121]: _type = "Task" [ 908.113387] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.124244] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106716, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.272776] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106714, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.299717] env[65121]: DEBUG nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 908.328379] env[65121]: DEBUG oslo_concurrency.lockutils [req-6d6560d6-2498-4556-8239-9cd5c5120f68 req-00337efb-5d94-409b-b524-11f7a275c0b9 service nova] Releasing lock "refresh_cache-df5abcb5-583f-4b28-a074-3a3221d74d87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.337928] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 908.338191] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 908.338340] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 908.339155] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 908.339155] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 908.339155] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 908.339155] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.339155] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 908.339363] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 908.339421] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 908.339584] env[65121]: DEBUG nova.virt.hardware [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 908.340598] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f6d4e5-4995-4748-96fd-d47007b08b68 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.358937] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106715, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.360577] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f56498-c380-47da-bbb8-8b2fb40b9efa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.474894] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: bad21c0b-6534-46b5-97a3-a26e5ecb5c96] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 908.569025] env[65121]: DEBUG nova.scheduler.client.report [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 908.628278] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106716, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.769027] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106714, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595169} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.770141] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.770141] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.770399] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20d24ab1-c8d7-4d54-bf71-ebcd141198c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.778869] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 908.778869] env[65121]: value = "task-5106717" [ 908.778869] env[65121]: _type = "Task" [ 908.778869] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.790313] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106717, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.851308] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106715, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.809205} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.851586] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] df5abcb5-583f-4b28-a074-3a3221d74d87/df5abcb5-583f-4b28-a074-3a3221d74d87.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.851794] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.852062] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4026805c-b8b8-4563-b56c-4f5ca0f2c9c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.864916] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 908.864916] env[65121]: value = "task-5106718" [ 908.864916] env[65121]: _type = "Task" [ 908.864916] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.875523] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106718, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.978878] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: aa9b6708-c53c-4117-9b75-9d506f393395] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 909.077598] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.833s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.077853] env[65121]: INFO nova.compute.manager [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Migrating [ 909.089638] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.852s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.091490] env[65121]: INFO nova.compute.claims [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.126858] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106716, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.296462] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106717, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072316} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.296959] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.298689] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a2b922-119d-45e0-86cc-0b244424fdeb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.322919] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.323330] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2b6c37d-4085-4792-89f6-26bba888e99a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.348032] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 909.348032] env[65121]: value = "task-5106719" [ 909.348032] env[65121]: _type = "Task" [ 909.348032] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.355227] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106719, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.375907] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106718, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.251464} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.376204] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.377478] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f876678e-1231-4113-841f-06a7198f9f2d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.402724] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] df5abcb5-583f-4b28-a074-3a3221d74d87/df5abcb5-583f-4b28-a074-3a3221d74d87.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.403054] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4ff8b96-9cea-4464-b578-6f9841ff69ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.426176] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 909.426176] env[65121]: value = "task-5106720" [ 909.426176] env[65121]: _type = "Task" [ 909.426176] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.435348] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106720, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.482226] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 2700ee2c-cc57-40e0-8069-f1b7ad2df7b4] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 909.602915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.602915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 909.602915] env[65121]: DEBUG nova.network.neutron [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 909.631298] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106716, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.796318] env[65121]: DEBUG nova.network.neutron [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Successfully updated port: c2c37a16-3107-4510-9c9d-17bf130945ff {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 909.866098] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.937278] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106720, 'name': ReconfigVM_Task, 'duration_secs': 0.386464} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.937679] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Reconfigured VM instance instance-00000045 to attach disk [datastore1] df5abcb5-583f-4b28-a074-3a3221d74d87/df5abcb5-583f-4b28-a074-3a3221d74d87.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.938364] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fc3f9f2-b890-40ad-a0c4-6a309d19a282 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.948634] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 909.948634] env[65121]: value = "task-5106721" [ 909.948634] env[65121]: _type = "Task" [ 909.948634] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.960823] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106721, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.985641] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: a43879c7-7378-4700-9581-8daabdafe5b2] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 910.106471] env[65121]: WARNING neutronclient.v2_0.client [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.107309] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.107737] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.127217] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106716, 'name': CreateVM_Task, 'duration_secs': 1.720344} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.127529] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.127999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.128172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 910.128549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 910.128833] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e632fa57-a857-49c8-ace3-069f48f6ec01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.134714] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 910.134714] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529794b3-5759-6938-d46a-f12528ed2ac6" [ 910.134714] env[65121]: _type = "Task" [ 910.134714] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.151498] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529794b3-5759-6938-d46a-f12528ed2ac6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.301784] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-600d5554-f52a-48ca-941d-1a755d086823" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.301871] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-600d5554-f52a-48ca-941d-1a755d086823" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 910.302065] env[65121]: DEBUG nova.network.neutron [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 910.364578] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106719, 'name': ReconfigVM_Task, 'duration_secs': 0.857711} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.365095] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.365824] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3709f93e-5eec-4ac8-bc15-00b80c4d3642 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.378620] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 910.378620] env[65121]: value = "task-5106722" [ 910.378620] env[65121]: _type = "Task" [ 910.378620] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.391927] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106722, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.463629] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106721, 'name': Rename_Task, 'duration_secs': 0.288668} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.463629] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.463629] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc1c8a62-8328-4e72-99a1-45aa1052c4c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.473036] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 910.473036] env[65121]: value = "task-5106723" [ 910.473036] env[65121]: _type = "Task" [ 910.473036] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.482483] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106723, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.493952] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 0151ed85-bba4-443a-b18f-58272c973e1e] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 910.588017] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.588534] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.635919] env[65121]: DEBUG nova.compute.manager [req-83ee0c63-0e7e-40ba-8ffc-88e86a4c6b06 req-2243f916-658b-447e-a4b9-a0fdc5385e2b service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Received event network-vif-plugged-c2c37a16-3107-4510-9c9d-17bf130945ff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 910.636244] env[65121]: DEBUG oslo_concurrency.lockutils [req-83ee0c63-0e7e-40ba-8ffc-88e86a4c6b06 req-2243f916-658b-447e-a4b9-a0fdc5385e2b service nova] Acquiring lock "600d5554-f52a-48ca-941d-1a755d086823-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.636391] env[65121]: DEBUG oslo_concurrency.lockutils [req-83ee0c63-0e7e-40ba-8ffc-88e86a4c6b06 req-2243f916-658b-447e-a4b9-a0fdc5385e2b service nova] Lock "600d5554-f52a-48ca-941d-1a755d086823-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.636877] env[65121]: DEBUG oslo_concurrency.lockutils [req-83ee0c63-0e7e-40ba-8ffc-88e86a4c6b06 req-2243f916-658b-447e-a4b9-a0fdc5385e2b service nova] Lock "600d5554-f52a-48ca-941d-1a755d086823-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.636877] env[65121]: DEBUG nova.compute.manager [req-83ee0c63-0e7e-40ba-8ffc-88e86a4c6b06 req-2243f916-658b-447e-a4b9-a0fdc5385e2b service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] No waiting events found dispatching network-vif-plugged-c2c37a16-3107-4510-9c9d-17bf130945ff {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 910.636952] env[65121]: WARNING nova.compute.manager [req-83ee0c63-0e7e-40ba-8ffc-88e86a4c6b06 req-2243f916-658b-447e-a4b9-a0fdc5385e2b service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Received unexpected event network-vif-plugged-c2c37a16-3107-4510-9c9d-17bf130945ff for instance with vm_state building and task_state spawning. [ 910.654716] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529794b3-5759-6938-d46a-f12528ed2ac6, 'name': SearchDatastore_Task, 'duration_secs': 0.016695} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.655075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 910.656075] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.656075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.656075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 910.656075] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.656469] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-147906d1-1a63-4579-b0a7-9a1608335255 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.672560] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.672759] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 910.674726] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1366fdd1-eee5-4527-8b1d-10f42c483187 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.680744] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 910.680744] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52aec9ab-0f68-7763-6ddb-a72a8e715dd9" [ 910.680744] env[65121]: _type = "Task" [ 910.680744] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.694352] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52aec9ab-0f68-7763-6ddb-a72a8e715dd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.702722] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7160d64c-b988-48ef-b158-0825396a9b3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.714685] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306d357f-3a74-4619-85c8-d325cf2cd97e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.723984] env[65121]: WARNING neutronclient.v2_0.client [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.725107] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.725819] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.776195] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c545a4f-2abc-4773-87f6-a5a17ba336da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.785197] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d066262c-2e9d-4e3c-9277-bd4cbca540a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.804454] env[65121]: DEBUG nova.compute.provider_tree [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.809715] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.809715] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.865054] env[65121]: DEBUG nova.network.neutron [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 910.893037] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106722, 'name': Rename_Task, 'duration_secs': 0.337822} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.893611] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.893705] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63e968f7-89a1-4375-9ed8-9129b2b5eace {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.896314] env[65121]: DEBUG nova.network.neutron [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 910.906577] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 910.906577] env[65121]: value = "task-5106724" [ 910.906577] env[65121]: _type = "Task" [ 910.906577] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.916794] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106724, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.942176] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.942742] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.985333] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106723, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.997077] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 20396b81-2041-4bf2-85c4-8ea66c838ec4] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 911.035709] env[65121]: WARNING neutronclient.v2_0.client [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.037200] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.037200] env[65121]: WARNING openstack [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.142675] env[65121]: DEBUG nova.network.neutron [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Updating instance_info_cache with network_info: [{"id": "c2c37a16-3107-4510-9c9d-17bf130945ff", "address": "fa:16:3e:92:20:ff", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2c37a16-31", "ovs_interfaceid": "c2c37a16-3107-4510-9c9d-17bf130945ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 911.194794] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52aec9ab-0f68-7763-6ddb-a72a8e715dd9, 'name': SearchDatastore_Task, 'duration_secs': 0.017863} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.196202] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-498b40b0-9182-4ae1-ac6c-27520bf8e925 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.202760] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 911.202760] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e8f776-b2fd-4ce8-29f6-8686bffb6aaf" [ 911.202760] env[65121]: _type = "Task" [ 911.202760] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.213841] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e8f776-b2fd-4ce8-29f6-8686bffb6aaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.307700] env[65121]: DEBUG nova.scheduler.client.report [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 911.367612] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.418523] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106724, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.487406] env[65121]: DEBUG oslo_vmware.api [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106723, 'name': PowerOnVM_Task, 'duration_secs': 0.520767} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.487750] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.487971] env[65121]: INFO nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Took 8.62 seconds to spawn the instance on the hypervisor. [ 911.488194] env[65121]: DEBUG nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 911.489121] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0d1da0-1cf0-425c-b75f-a3189d73aade {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.499688] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: db737254-91fd-44bc-bf01-c864227cdf07] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 911.548459] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.548715] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.548920] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.549103] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.549266] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.551652] env[65121]: INFO nova.compute.manager [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Terminating instance [ 911.644589] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-600d5554-f52a-48ca-941d-1a755d086823" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.645091] env[65121]: DEBUG nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Instance network_info: |[{"id": "c2c37a16-3107-4510-9c9d-17bf130945ff", "address": "fa:16:3e:92:20:ff", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2c37a16-31", "ovs_interfaceid": "c2c37a16-3107-4510-9c9d-17bf130945ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 911.645640] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:20:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2c37a16-3107-4510-9c9d-17bf130945ff', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.665079] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 911.665504] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.665857] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c66b952-17d3-49a2-90ee-053925cdb7de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.697288] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.697288] env[65121]: value = "task-5106725" [ 911.697288] env[65121]: _type = "Task" [ 911.697288] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.706308] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106725, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.716269] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e8f776-b2fd-4ce8-29f6-8686bffb6aaf, 'name': SearchDatastore_Task, 'duration_secs': 0.01244} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.716269] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.716398] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 911.716565] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-538a8f1a-d39b-4262-a228-68717177c8e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.724030] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 911.724030] env[65121]: value = "task-5106726" [ 911.724030] env[65121]: _type = "Task" [ 911.724030] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.737130] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106726, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.817487] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.727s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.818111] env[65121]: DEBUG nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 911.821685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.669s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.821964] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.825030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.149s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.827344] env[65121]: INFO nova.compute.claims [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.854474] env[65121]: INFO nova.scheduler.client.report [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Deleted allocations for instance c2f9fa49-4273-4885-b59d-0635ed739d65 [ 911.923693] env[65121]: DEBUG oslo_vmware.api [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106724, 'name': PowerOnVM_Task, 'duration_secs': 0.769457} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.923693] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.923693] env[65121]: INFO nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Took 6.43 seconds to spawn the instance on the hypervisor. [ 911.923693] env[65121]: DEBUG nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 911.924539] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd37413-a040-4e71-b5f3-a25a3bf45225 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.007925] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 912.008377] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Cleaning up deleted instances with incomplete migration {{(pid=65121) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 912.012754] env[65121]: INFO nova.compute.manager [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Took 56.10 seconds to build instance. [ 912.055946] env[65121]: DEBUG nova.compute.manager [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 912.056336] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.057594] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda7a7a5-5acf-46ab-88bc-8b98268f1966 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.069051] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.069429] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f0b1aaf-8b4b-49ec-b209-b0e272fe6488 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.079930] env[65121]: DEBUG oslo_vmware.api [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 912.079930] env[65121]: value = "task-5106727" [ 912.079930] env[65121]: _type = "Task" [ 912.079930] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.095218] env[65121]: DEBUG oslo_vmware.api [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106727, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.217535] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106725, 'name': CreateVM_Task, 'duration_secs': 0.409145} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.217535] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.219234] env[65121]: WARNING neutronclient.v2_0.client [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.219234] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.219234] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.219234] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 912.219234] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d75ffec-2c28-4a1c-be24-ccfcc839c75b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.230022] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 912.230022] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e9e683-a50d-f33a-33da-119744ad62ee" [ 912.230022] env[65121]: _type = "Task" [ 912.230022] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.238950] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106726, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.243389] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e9e683-a50d-f33a-33da-119744ad62ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.334981] env[65121]: DEBUG nova.compute.utils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 912.343206] env[65121]: DEBUG nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 912.343206] env[65121]: DEBUG nova.network.neutron [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 912.343392] env[65121]: WARNING neutronclient.v2_0.client [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.343761] env[65121]: WARNING neutronclient.v2_0.client [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.345090] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.346306] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.365039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7d3d4d35-b9cf-4158-adbc-2a45b3ebceaf tempest-ListServerFiltersTestJSON-516348507 tempest-ListServerFiltersTestJSON-516348507-project-member] Lock "c2f9fa49-4273-4885-b59d-0635ed739d65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.798s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.375778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 912.375778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.391381] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "293f93f2-c01d-42c8-b1a7-3056805c77de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 912.391381] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.391520] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 912.391658] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.391829] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.395417] env[65121]: INFO nova.compute.manager [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Terminating instance [ 912.418268] env[65121]: DEBUG nova.policy [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e773a3343a448f5acd3f08005e8ed45', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '40aee06073b04c25bc7b38fbd09856e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 912.448238] env[65121]: INFO nova.compute.manager [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Took 51.23 seconds to build instance. [ 912.516182] env[65121]: DEBUG oslo_concurrency.lockutils [None req-92bf199e-d56e-47a7-816f-4cee12f78117 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.037s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.516573] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 912.593956] env[65121]: DEBUG oslo_vmware.api [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106727, 'name': PowerOffVM_Task, 'duration_secs': 0.285421} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.594276] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.594538] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 912.594842] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9b3cfcd-9a2a-4697-afdd-f960db1a411c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.675176] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.678019] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.678019] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Deleting the datastore file [datastore2] b22ccce7-c54a-4577-9de0-1fd9c10cd189 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.678019] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd1c7a4c-5451-47bc-b676-6590de4f2c35 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.688931] env[65121]: DEBUG oslo_vmware.api [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for the task: (returnval){ [ 912.688931] env[65121]: value = "task-5106729" [ 912.688931] env[65121]: _type = "Task" [ 912.688931] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.695514] env[65121]: DEBUG oslo_vmware.api [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.738660] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106726, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581325} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.739350] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.739609] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.739886] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4811f4c9-db9a-4a01-9f5e-6be2ecbcc54d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.745621] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e9e683-a50d-f33a-33da-119744ad62ee, 'name': SearchDatastore_Task, 'duration_secs': 0.049586} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.746259] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 912.746536] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.746719] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.746857] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.747052] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.747293] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd9fe903-fbab-4a7d-a444-257ada705138 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.751573] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 912.751573] env[65121]: value = "task-5106730" [ 912.751573] env[65121]: _type = "Task" [ 912.751573] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.756834] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.756990] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.758228] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-144a7b2b-92d2-4745-9da6-7b540e8ae857 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.765670] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106730, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.769102] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 912.769102] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f44865-6669-972f-b01d-4cea15e8ca80" [ 912.769102] env[65121]: _type = "Task" [ 912.769102] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.778342] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f44865-6669-972f-b01d-4cea15e8ca80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.843410] env[65121]: DEBUG nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 912.881522] env[65121]: DEBUG nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 912.892693] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4708428e-1d12-4465-a028-a277de81c549 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.904301] env[65121]: DEBUG nova.compute.manager [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 912.904769] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.906457] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0ea4f6-1363-4f68-aafa-32f7c468913b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.925938] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance '7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 912.940035] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.940401] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-958d5149-9b5a-4e32-8957-6cbee661f9be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.948298] env[65121]: DEBUG oslo_vmware.api [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 912.948298] env[65121]: value = "task-5106731" [ 912.948298] env[65121]: _type = "Task" [ 912.948298] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.954147] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5d223c99-e8c5-449d-812f-4e96488fa87a tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "2e676b64-181e-4fee-8120-05cb49b5bb4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.789s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.963248] env[65121]: DEBUG oslo_vmware.api [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.975738] env[65121]: DEBUG nova.network.neutron [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Successfully created port: 265542f8-48b6-48e7-b0c5-2d371a74f6da {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 913.202940] env[65121]: DEBUG oslo_vmware.api [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.265478] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106730, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071613} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.265797] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.266637] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a725c2-60b7-4dfc-8300-c45d043c1d03 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.290643] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.294306] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86777e2a-2771-43e1-8297-cf2a6f127159 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.315776] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f44865-6669-972f-b01d-4cea15e8ca80, 'name': SearchDatastore_Task, 'duration_secs': 0.012913} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.317276] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dc78d75-ea2d-4141-bd89-9b226dcd0e75 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.323918] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 913.323918] env[65121]: value = "task-5106732" [ 913.323918] env[65121]: _type = "Task" [ 913.323918] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.324175] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 913.324175] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524027a7-fb1d-c567-d0d7-fabef7c20e79" [ 913.324175] env[65121]: _type = "Task" [ 913.324175] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.338689] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524027a7-fb1d-c567-d0d7-fabef7c20e79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.342120] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106732, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.407068] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.432375] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.433530] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-300c9529-8e43-44b7-9dfa-e75391631f15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.443511] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 913.443511] env[65121]: value = "task-5106733" [ 913.443511] env[65121]: _type = "Task" [ 913.443511] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.456492] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 913.456945] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance '7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 913.466403] env[65121]: DEBUG oslo_vmware.api [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106731, 'name': PowerOffVM_Task, 'duration_secs': 0.313911} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.469441] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.469441] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.469441] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a3314de-9449-4431-abef-8deb88f7f061 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.552490] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 913.552663] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 913.553018] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Deleting the datastore file [datastore1] 293f93f2-c01d-42c8-b1a7-3056805c77de {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.553579] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-624d69b6-1be8-489a-8d2a-cdbd7bd1c47b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.558445] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de2f3f9-7c42-486b-b63d-a86fd79634ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.572497] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d60f5d9-e595-4d85-99a1-1528d7612b3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.576932] env[65121]: DEBUG oslo_vmware.api [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 913.576932] env[65121]: value = "task-5106735" [ 913.576932] env[65121]: _type = "Task" [ 913.576932] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.615134] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8dea86-6ecd-4420-9f9d-eb11f92fedf5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.622530] env[65121]: DEBUG oslo_vmware.api [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106735, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.628798] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5e4097-9747-4e83-9376-753d35624913 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.647819] env[65121]: DEBUG nova.compute.provider_tree [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.696363] env[65121]: DEBUG nova.compute.manager [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Received event network-changed-c2c37a16-3107-4510-9c9d-17bf130945ff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 913.696604] env[65121]: DEBUG nova.compute.manager [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Refreshing instance network info cache due to event network-changed-c2c37a16-3107-4510-9c9d-17bf130945ff. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 913.696938] env[65121]: DEBUG oslo_concurrency.lockutils [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Acquiring lock "refresh_cache-600d5554-f52a-48ca-941d-1a755d086823" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.697079] env[65121]: DEBUG oslo_concurrency.lockutils [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Acquired lock "refresh_cache-600d5554-f52a-48ca-941d-1a755d086823" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 913.697267] env[65121]: DEBUG nova.network.neutron [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Refreshing network info cache for port c2c37a16-3107-4510-9c9d-17bf130945ff {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 913.706034] env[65121]: DEBUG oslo_vmware.api [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Task: {'id': task-5106729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.527597} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.707033] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.707151] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.707438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.707652] env[65121]: INFO nova.compute.manager [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Took 1.65 seconds to destroy the instance on the hypervisor. [ 913.708161] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 913.708495] env[65121]: DEBUG nova.compute.manager [-] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 913.708613] env[65121]: DEBUG nova.network.neutron [-] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 913.708894] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.709509] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.709800] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.797736] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.838642] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106732, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.841099] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524027a7-fb1d-c567-d0d7-fabef7c20e79, 'name': SearchDatastore_Task, 'duration_secs': 0.026757} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.841430] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 913.841707] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 913.842370] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1e7a339-cedd-4027-b82e-a74089f2c5e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.850687] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 913.850687] env[65121]: value = "task-5106736" [ 913.850687] env[65121]: _type = "Task" [ 913.850687] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.855282] env[65121]: DEBUG nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 913.863368] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.885892] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='16bb4fa3a061fe051cfc0a98e37c53dc',container_format='bare',created_at=2025-12-12T14:24:19Z,direct_url=,disk_format='vmdk',id=e193384d-79f8-40cc-874e-60556f28628f,min_disk=1,min_ram=0,name='tempest-test-snap-928740552',owner='40aee06073b04c25bc7b38fbd09856e9',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-12-12T14:24:38Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 913.886176] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 913.886326] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 913.886607] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 913.886692] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 913.886828] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 913.887051] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.887207] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 913.887373] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 913.887528] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 913.887773] env[65121]: DEBUG nova.virt.hardware [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 913.888623] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29512c2-32df-48e3-8b82-fef9297747aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.897628] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c1e97f-75cf-4167-abff-4d70232f9c12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.969711] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 913.969965] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 913.970219] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 913.970415] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 913.970555] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 913.970698] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 913.970892] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.971100] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 913.971476] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 913.971743] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 913.971955] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 913.978486] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e7f341b-ae66-47c4-aab8-1e43eab566e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.996038] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 913.996038] env[65121]: value = "task-5106737" [ 913.996038] env[65121]: _type = "Task" [ 913.996038] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.007960] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106737, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.089164] env[65121]: DEBUG oslo_vmware.api [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106735, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207442} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.089427] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.090080] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.090285] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.091246] env[65121]: INFO nova.compute.manager [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Took 1.19 seconds to destroy the instance on the hypervisor. [ 914.091246] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 914.091246] env[65121]: DEBUG nova.compute.manager [-] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 914.091246] env[65121]: DEBUG nova.network.neutron [-] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 914.091513] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.092163] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.092361] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.151972] env[65121]: DEBUG nova.scheduler.client.report [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 914.201672] env[65121]: WARNING neutronclient.v2_0.client [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.202385] env[65121]: WARNING openstack [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.203204] env[65121]: WARNING openstack [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.344593] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106732, 'name': ReconfigVM_Task, 'duration_secs': 0.603303} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.345380] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec/30b01b2b-2cbf-4c34-86cd-529a95a9c3ec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.346785] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-801fd2a8-622d-4937-9eaf-997c32c6d1de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.360582] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 914.360582] env[65121]: value = "task-5106738" [ 914.360582] env[65121]: _type = "Task" [ 914.360582] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.371172] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106736, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.372777] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.387238] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106738, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.510239] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106737, 'name': ReconfigVM_Task, 'duration_secs': 0.345127} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.510572] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance '7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 914.659228] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.834s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.659839] env[65121]: DEBUG nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 914.663506] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.218s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.663779] env[65121]: DEBUG nova.objects.instance [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 914.729946] env[65121]: DEBUG nova.network.neutron [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Successfully updated port: 265542f8-48b6-48e7-b0c5-2d371a74f6da {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 914.792682] env[65121]: DEBUG nova.network.neutron [-] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.863781] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106736, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.873375] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106738, 'name': Rename_Task, 'duration_secs': 0.195223} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.873707] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.876104] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf9981c1-f3a5-4cbe-99c4-6229ecb938c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.885379] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 914.885379] env[65121]: value = "task-5106739" [ 914.885379] env[65121]: _type = "Task" [ 914.885379] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.892705] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106739, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.946480] env[65121]: WARNING openstack [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.946480] env[65121]: WARNING openstack [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.020706] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 915.020980] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 915.021449] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 915.021449] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 915.021608] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 915.021805] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 915.022050] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.022692] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 915.022692] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 915.022692] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 915.022874] env[65121]: DEBUG nova.virt.hardware [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 915.028427] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Reconfiguring VM instance instance-0000001d to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 915.029007] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5997a87-fc9a-4855-9041-fe2329506a8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.049586] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 915.049586] env[65121]: value = "task-5106740" [ 915.049586] env[65121]: _type = "Task" [ 915.049586] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.059421] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106740, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.079599] env[65121]: WARNING neutronclient.v2_0.client [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.080176] env[65121]: WARNING openstack [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.080529] env[65121]: WARNING openstack [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.172189] env[65121]: DEBUG nova.compute.utils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 915.180126] env[65121]: DEBUG nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 915.180126] env[65121]: DEBUG nova.network.neutron [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 915.180126] env[65121]: WARNING neutronclient.v2_0.client [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.180126] env[65121]: WARNING neutronclient.v2_0.client [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.180126] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.180655] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.196054] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6d723ac4-ca7b-4896-bf95-afd9de9a0a63 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.533s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.198388] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.710s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.204912] env[65121]: INFO nova.compute.claims [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.211139] env[65121]: DEBUG nova.network.neutron [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Updated VIF entry in instance network info cache for port c2c37a16-3107-4510-9c9d-17bf130945ff. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 915.211139] env[65121]: DEBUG nova.network.neutron [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Updating instance_info_cache with network_info: [{"id": "c2c37a16-3107-4510-9c9d-17bf130945ff", "address": "fa:16:3e:92:20:ff", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2c37a16-31", "ovs_interfaceid": "c2c37a16-3107-4510-9c9d-17bf130945ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.234405] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "refresh_cache-b2ebce1d-a8d8-4a71-997b-39d09ca373fe" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.234548] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "refresh_cache-b2ebce1d-a8d8-4a71-997b-39d09ca373fe" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 915.234845] env[65121]: DEBUG nova.network.neutron [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 915.260389] env[65121]: DEBUG nova.policy [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1c51fc43b7f44ce9b023d8cb769e029', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89110618b6ef4ce68b2ca53e7e168139', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 915.264284] env[65121]: DEBUG oslo_concurrency.lockutils [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "df5abcb5-583f-4b28-a074-3a3221d74d87" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.264536] env[65121]: DEBUG oslo_concurrency.lockutils [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.302750] env[65121]: INFO nova.compute.manager [-] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Took 1.59 seconds to deallocate network for instance. [ 915.368393] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106736, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.394425] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106739, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.433284] env[65121]: DEBUG nova.network.neutron [-] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.561132] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.624019] env[65121]: DEBUG nova.network.neutron [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Successfully created port: 98acae83-de55-4870-8650-eb502346b7e4 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 915.694630] env[65121]: DEBUG nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 915.713677] env[65121]: DEBUG oslo_concurrency.lockutils [req-e5e51a6c-f74c-4ec0-9270-5f252f6bb6fb req-88e10eb7-3ae8-42fd-a27a-4d7aaf5659d3 service nova] Releasing lock "refresh_cache-600d5554-f52a-48ca-941d-1a755d086823" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 915.741032] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.741032] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.769655] env[65121]: DEBUG nova.compute.utils [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 915.812460] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.866941] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106736, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.721435} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.866941] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.866941] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.866941] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07e5460d-b579-464a-9b2d-f0bc5ac5e760 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.871641] env[65121]: DEBUG nova.network.neutron [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 915.875517] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 915.875517] env[65121]: value = "task-5106741" [ 915.875517] env[65121]: _type = "Task" [ 915.875517] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.886168] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106741, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.895536] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106739, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.939830] env[65121]: INFO nova.compute.manager [-] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Took 1.85 seconds to deallocate network for instance. [ 916.062163] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.274190] env[65121]: DEBUG oslo_concurrency.lockutils [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.386400] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106741, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076013} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.391806] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.392719] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a8056f-9c66-49db-b8ec-cf717d00989c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.401110] env[65121]: DEBUG oslo_vmware.api [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106739, 'name': PowerOnVM_Task, 'duration_secs': 1.125889} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.410583] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.410794] env[65121]: DEBUG nova.compute.manager [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 916.420711] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.424036] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aab1c67-b188-48f6-95af-34f87d6f6631 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.426528] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b695973a-8e8f-46ae-a1d7-a1fd0f3fb6b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.447956] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.453205] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 916.453205] env[65121]: value = "task-5106742" [ 916.453205] env[65121]: _type = "Task" [ 916.453205] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.466205] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106742, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.564488] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106740, 'name': ReconfigVM_Task, 'duration_secs': 1.083417} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.567559] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Reconfigured VM instance instance-0000001d to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 916.569657] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9f1053-2276-40bd-89ee-eafb4d7cb037 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.595046] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c/7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.597339] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7014e2b-4911-462c-91b7-6bde84574e0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.620757] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 916.620757] env[65121]: value = "task-5106743" [ 916.620757] env[65121]: _type = "Task" [ 916.620757] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.632060] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106743, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.705532] env[65121]: DEBUG nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 916.733365] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 916.733610] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 916.733784] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 916.734138] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 916.734227] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 916.734344] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 916.734542] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.734694] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 916.734850] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 916.735011] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 916.735259] env[65121]: DEBUG nova.virt.hardware [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 916.737274] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033a347a-72d4-420d-a98d-db08a6d4e182 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.746246] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371889cb-1f8a-4b4d-822b-2cce4f1e345f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.782251] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.782628] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.791196] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b3aaa4-cdf8-4dbb-ac83-2b6f5fccc688 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.800340] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cce48e-e6c8-4204-835e-9245888b3064 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.838385] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a553b37-6b38-479f-b532-491eb578c316 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.850879] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ec6fc6-9bec-41cf-a36c-54754a6f2272 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.866705] env[65121]: DEBUG nova.compute.provider_tree [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.963477] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.969021] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106742, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.132637] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106743, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.212961] env[65121]: DEBUG nova.network.neutron [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Successfully updated port: 98acae83-de55-4870-8650-eb502346b7e4 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 917.247719] env[65121]: WARNING neutronclient.v2_0.client [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.248648] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.249025] env[65121]: WARNING openstack [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.348499] env[65121]: DEBUG oslo_concurrency.lockutils [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "df5abcb5-583f-4b28-a074-3a3221d74d87" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.349034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.349034] env[65121]: INFO nova.compute.manager [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Attaching volume 1d017003-f486-4c1b-aa9a-dfd8d2526776 to /dev/sdb [ 917.371076] env[65121]: DEBUG nova.scheduler.client.report [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 917.399087] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d577350-ca7a-4f64-916c-59a3ac896478 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.408704] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7496d141-9b0d-4081-b0e2-20039b98dc09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.421560] env[65121]: DEBUG nova.compute.manager [req-5b0fa67b-e649-44df-a1c7-ea873fd48e85 req-0becef35-36cc-498b-99d5-e0261ed97556 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Received event network-vif-plugged-265542f8-48b6-48e7-b0c5-2d371a74f6da {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 917.421800] env[65121]: DEBUG oslo_concurrency.lockutils [req-5b0fa67b-e649-44df-a1c7-ea873fd48e85 req-0becef35-36cc-498b-99d5-e0261ed97556 service nova] Acquiring lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.422016] env[65121]: DEBUG oslo_concurrency.lockutils [req-5b0fa67b-e649-44df-a1c7-ea873fd48e85 req-0becef35-36cc-498b-99d5-e0261ed97556 service nova] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.422196] env[65121]: DEBUG oslo_concurrency.lockutils [req-5b0fa67b-e649-44df-a1c7-ea873fd48e85 req-0becef35-36cc-498b-99d5-e0261ed97556 service nova] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.422358] env[65121]: DEBUG nova.compute.manager [req-5b0fa67b-e649-44df-a1c7-ea873fd48e85 req-0becef35-36cc-498b-99d5-e0261ed97556 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] No waiting events found dispatching network-vif-plugged-265542f8-48b6-48e7-b0c5-2d371a74f6da {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 917.422512] env[65121]: WARNING nova.compute.manager [req-5b0fa67b-e649-44df-a1c7-ea873fd48e85 req-0becef35-36cc-498b-99d5-e0261ed97556 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Received unexpected event network-vif-plugged-265542f8-48b6-48e7-b0c5-2d371a74f6da for instance with vm_state building and task_state spawning. [ 917.427351] env[65121]: DEBUG nova.virt.block_device [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Updating existing volume attachment record: 804ffffb-8ef5-43ab-98a0-72ace75eadd7 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 917.438543] env[65121]: DEBUG nova.network.neutron [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Updating instance_info_cache with network_info: [{"id": "265542f8-48b6-48e7-b0c5-2d371a74f6da", "address": "fa:16:3e:c3:6b:49", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap265542f8-48", "ovs_interfaceid": "265542f8-48b6-48e7-b0c5-2d371a74f6da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 917.473028] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106742, 'name': ReconfigVM_Task, 'duration_secs': 0.731349} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.473028] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.473028] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-751349ad-034e-4281-aae7-ce0e4cc2c37f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.481026] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 917.481026] env[65121]: value = "task-5106744" [ 917.481026] env[65121]: _type = "Task" [ 917.481026] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.489926] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106744, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.641544] env[65121]: DEBUG oslo_vmware.api [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106743, 'name': ReconfigVM_Task, 'duration_secs': 0.968645} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.641714] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Reconfigured VM instance instance-0000001d to attach disk [datastore1] 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c/7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.641985] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance '7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 917.722368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "refresh_cache-c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.722368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "refresh_cache-c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 917.722368] env[65121]: DEBUG nova.network.neutron [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 917.802108] env[65121]: DEBUG nova.compute.manager [req-4f2e5353-bc46-4bdc-9814-0c92032b3db0 req-521fdbe0-2fdd-49d2-9a67-8253d6b39525 service nova] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Received event network-vif-deleted-2e0f93ed-17fe-4ad2-aadd-13b7bdac954c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 917.802320] env[65121]: DEBUG nova.compute.manager [req-4f2e5353-bc46-4bdc-9814-0c92032b3db0 req-521fdbe0-2fdd-49d2-9a67-8253d6b39525 service nova] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Received event network-vif-deleted-79bfb570-6ed2-4c0f-83e1-1f2cfb088e37 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 917.878019] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.678s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.878019] env[65121]: DEBUG nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 917.879865] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.400s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.880214] env[65121]: DEBUG nova.objects.instance [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 93721179-407a-43d9-b0bf-157433a09519] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 917.947191] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "refresh_cache-b2ebce1d-a8d8-4a71-997b-39d09ca373fe" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.947191] env[65121]: DEBUG nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Instance network_info: |[{"id": "265542f8-48b6-48e7-b0c5-2d371a74f6da", "address": "fa:16:3e:c3:6b:49", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap265542f8-48", "ovs_interfaceid": "265542f8-48b6-48e7-b0c5-2d371a74f6da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 917.947403] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:6b:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '265542f8-48b6-48e7-b0c5-2d371a74f6da', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.959920] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 917.960554] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.960795] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01a07bab-a346-41eb-bb54-d1d56d824e53 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.986217] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.986217] env[65121]: value = "task-5106748" [ 917.986217] env[65121]: _type = "Task" [ 917.986217] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.994438] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106744, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.004991] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106748, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.153025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee1f154-01c4-4aed-8927-f52054aa3eb5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.180111] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c2adfe-922a-4613-a8e6-48af3a9e2930 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.209033] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance '7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 918.227088] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.227571] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.278878] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 918.279193] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.301344] env[65121]: DEBUG nova.network.neutron [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 918.329600] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.329992] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.387305] env[65121]: DEBUG nova.compute.utils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 918.392210] env[65121]: DEBUG nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 918.393109] env[65121]: DEBUG nova.network.neutron [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 918.393318] env[65121]: WARNING neutronclient.v2_0.client [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.393927] env[65121]: WARNING neutronclient.v2_0.client [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.394635] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.395888] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.408672] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0048f020-3d89-48a6-85dd-2d12cec08934 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.529s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 918.409689] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.810s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.415447] env[65121]: INFO nova.compute.claims [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.430878] env[65121]: WARNING neutronclient.v2_0.client [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.431570] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.432100] env[65121]: WARNING openstack [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.480138] env[65121]: DEBUG nova.policy [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bf232407a7346c98ea64ed673a179f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b35af4e2b6844a9081f34466661c83b2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 918.508538] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106744, 'name': Rename_Task, 'duration_secs': 0.859184} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.512758] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.513234] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106748, 'name': CreateVM_Task, 'duration_secs': 0.337391} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.513428] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86c10ea4-0c82-4c78-858c-35f86c740502 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.515616] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.516857] env[65121]: WARNING neutronclient.v2_0.client [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.517318] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.517484] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 918.517847] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 918.518780] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acd0ca3d-78f9-4983-a46f-82423258c8fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.524950] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 918.524950] env[65121]: value = "task-5106749" [ 918.524950] env[65121]: _type = "Task" [ 918.524950] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.525709] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 918.525709] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5229a277-e4da-2aad-f50d-15786391f358" [ 918.525709] env[65121]: _type = "Task" [ 918.525709] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.539087] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106749, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.548554] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5229a277-e4da-2aad-f50d-15786391f358, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.717070] env[65121]: WARNING neutronclient.v2_0.client [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.723067] env[65121]: DEBUG nova.network.neutron [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Updating instance_info_cache with network_info: [{"id": "98acae83-de55-4870-8650-eb502346b7e4", "address": "fa:16:3e:84:29:ab", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98acae83-de", "ovs_interfaceid": "98acae83-de55-4870-8650-eb502346b7e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 918.787314] env[65121]: DEBUG nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 918.895141] env[65121]: DEBUG nova.network.neutron [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Successfully created port: b2c8d776-fc6a-4893-a904-f33386dc0317 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 918.907335] env[65121]: DEBUG nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 918.932454] env[65121]: DEBUG nova.network.neutron [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Port 0e25d678-e865-4027-804f-77e9bb4cb31a binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 918.997581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 918.997888] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.001257] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.001549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.003s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.001755] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.006128] env[65121]: INFO nova.compute.manager [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Terminating instance [ 919.049988] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106749, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.050407] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.050599] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Processing image e193384d-79f8-40cc-874e-60556f28628f {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.050817] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f/e193384d-79f8-40cc-874e-60556f28628f.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.050956] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f/e193384d-79f8-40cc-874e-60556f28628f.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.051149] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.051431] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eae31243-b3ee-439f-9394-0a693ff42fe4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.062097] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.062354] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.063260] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1eb5160-528b-49a3-a2c9-fb554581c570 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.071040] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 919.071040] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d43b6f-7525-e14e-104a-a89924ae344a" [ 919.071040] env[65121]: _type = "Task" [ 919.071040] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.080435] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d43b6f-7525-e14e-104a-a89924ae344a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.225758] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "refresh_cache-c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.226200] env[65121]: DEBUG nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Instance network_info: |[{"id": "98acae83-de55-4870-8650-eb502346b7e4", "address": "fa:16:3e:84:29:ab", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98acae83-de", "ovs_interfaceid": "98acae83-de55-4870-8650-eb502346b7e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 919.226900] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:29:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '98acae83-de55-4870-8650-eb502346b7e4', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.237623] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 919.237877] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.238122] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35a8a9b7-3cc7-40be-a344-3dfcac339eef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.266442] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.266442] env[65121]: value = "task-5106750" [ 919.266442] env[65121]: _type = "Task" [ 919.266442] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.275501] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106750, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.319188] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.389993] env[65121]: INFO nova.compute.manager [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Rebuilding instance [ 919.460309] env[65121]: DEBUG nova.compute.manager [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 919.460309] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8383ccab-e971-4450-af28-e022ff8fc618 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.512854] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "refresh_cache-30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.512854] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "refresh_cache-30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.512854] env[65121]: DEBUG nova.network.neutron [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 919.545555] env[65121]: DEBUG oslo_vmware.api [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106749, 'name': PowerOnVM_Task, 'duration_secs': 0.712578} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.545814] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.546012] env[65121]: INFO nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Took 11.25 seconds to spawn the instance on the hypervisor. [ 919.546184] env[65121]: DEBUG nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 919.547769] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e9b08e-ca1b-4724-9c85-71957cc4f6e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.587732] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Preparing fetch location {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 919.588017] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Fetch image to [datastore2] OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38/OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38.vmdk {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 919.588239] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Downloading stream optimized image e193384d-79f8-40cc-874e-60556f28628f to [datastore2] OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38/OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38.vmdk on the data store datastore2 as vApp {{(pid=65121) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 919.588490] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Downloading image file data e193384d-79f8-40cc-874e-60556f28628f to the ESX as VM named 'OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38' {{(pid=65121) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 919.724999] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 919.724999] env[65121]: value = "resgroup-9" [ 919.724999] env[65121]: _type = "ResourcePool" [ 919.724999] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 919.725907] env[65121]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-e99f6adb-d87d-4bc9-925d-d3f93eec8f23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.752732] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease: (returnval){ [ 919.752732] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5217c1fc-c6a4-3533-d158-9aacd6dc6473" [ 919.752732] env[65121]: _type = "HttpNfcLease" [ 919.752732] env[65121]: } obtained for vApp import into resource pool (val){ [ 919.752732] env[65121]: value = "resgroup-9" [ 919.752732] env[65121]: _type = "ResourcePool" [ 919.752732] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 919.752732] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the lease: (returnval){ [ 919.752732] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5217c1fc-c6a4-3533-d158-9aacd6dc6473" [ 919.752732] env[65121]: _type = "HttpNfcLease" [ 919.752732] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 919.762347] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 919.762347] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5217c1fc-c6a4-3533-d158-9aacd6dc6473" [ 919.762347] env[65121]: _type = "HttpNfcLease" [ 919.762347] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 919.780092] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106750, 'name': CreateVM_Task, 'duration_secs': 0.433572} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.780328] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.780960] env[65121]: WARNING neutronclient.v2_0.client [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.781429] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.781848] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.781990] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 919.782302] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-434683a7-66be-42f8-af00-013c4bde1264 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.789156] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 919.789156] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528022dd-cb7c-844e-7b82-aa710399a9e3" [ 919.789156] env[65121]: _type = "Task" [ 919.789156] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.807484] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528022dd-cb7c-844e-7b82-aa710399a9e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.918634] env[65121]: DEBUG nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 919.965858] env[65121]: DEBUG nova.virt.hardware [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 919.965858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.966456] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.966456] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.968122] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4389983-c69c-4bd7-9323-0793efbd75f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.983686] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e80e8b-a6b9-4244-8ced-281d339339eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.015274] env[65121]: WARNING neutronclient.v2_0.client [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.016082] env[65121]: WARNING openstack [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.016428] env[65121]: WARNING openstack [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.058963] env[65121]: DEBUG nova.network.neutron [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 920.073058] env[65121]: INFO nova.compute.manager [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Took 55.52 seconds to build instance. [ 920.084145] env[65121]: DEBUG nova.compute.manager [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Received event network-changed-265542f8-48b6-48e7-b0c5-2d371a74f6da {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 920.084145] env[65121]: DEBUG nova.compute.manager [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Refreshing instance network info cache due to event network-changed-265542f8-48b6-48e7-b0c5-2d371a74f6da. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 920.084234] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Acquiring lock "refresh_cache-b2ebce1d-a8d8-4a71-997b-39d09ca373fe" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.085326] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Acquired lock "refresh_cache-b2ebce1d-a8d8-4a71-997b-39d09ca373fe" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.085326] env[65121]: DEBUG nova.network.neutron [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Refreshing network info cache for port 265542f8-48b6-48e7-b0c5-2d371a74f6da {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 920.104136] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef58ae9c-3fc3-464e-a164-26a8f8d3d44e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.113922] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326592e4-05f7-450f-8e24-c8c52fad08e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.152260] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152c2f4b-6ddd-4841-b145-c55c979bebd0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.163766] env[65121]: DEBUG nova.network.neutron [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 920.166530] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663a131b-d9c0-49bf-931d-ee0debc2b671 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.183417] env[65121]: DEBUG nova.compute.provider_tree [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.260096] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 920.260096] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5217c1fc-c6a4-3533-d158-9aacd6dc6473" [ 920.260096] env[65121]: _type = "HttpNfcLease" [ 920.260096] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 920.309678] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528022dd-cb7c-844e-7b82-aa710399a9e3, 'name': SearchDatastore_Task, 'duration_secs': 0.015287} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.310129] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 920.310494] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.310788] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.311012] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.311279] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.311659] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c018c97-f463-4b35-8c65-e2fc4b06dc0d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.323368] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.323900] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.324615] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95f9ea2d-0c4e-458a-8085-7d7e3d32f103 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.331516] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 920.331516] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c43a90-9a79-8e29-21c9-685f658b1685" [ 920.331516] env[65121]: _type = "Task" [ 920.331516] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.340700] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c43a90-9a79-8e29-21c9-685f658b1685, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.483499] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.484151] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bef65582-56f1-4a73-9e25-be3a6d56103f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.493142] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 920.493142] env[65121]: value = "task-5106753" [ 920.493142] env[65121]: _type = "Task" [ 920.493142] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.505048] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.579214] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f737321c-ed4f-4348-b336-b1c6b92c0fdd tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.602s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.592998] env[65121]: WARNING neutronclient.v2_0.client [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.594169] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.594583] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.618330] env[65121]: DEBUG nova.network.neutron [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Successfully updated port: b2c8d776-fc6a-4893-a904-f33386dc0317 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 920.672875] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "refresh_cache-30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 920.673687] env[65121]: DEBUG nova.compute.manager [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 920.674911] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.676088] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aaec10e-5aa9-4ff1-bc67-fadfd2fb890c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.685285] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.685536] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65115046-6691-4d11-b1c9-b4d2dff8ae2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.687908] env[65121]: DEBUG nova.scheduler.client.report [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 920.704029] env[65121]: DEBUG oslo_vmware.api [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 920.704029] env[65121]: value = "task-5106754" [ 920.704029] env[65121]: _type = "Task" [ 920.704029] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.716166] env[65121]: DEBUG oslo_vmware.api [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.725461] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.725892] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.735258] env[65121]: DEBUG nova.compute.manager [req-24b6397f-9b0f-49f2-bacf-2227b1135902 req-6ee00a30-1e4b-41c5-940d-5768013fdc8d service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Received event network-vif-plugged-b2c8d776-fc6a-4893-a904-f33386dc0317 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 920.735465] env[65121]: DEBUG oslo_concurrency.lockutils [req-24b6397f-9b0f-49f2-bacf-2227b1135902 req-6ee00a30-1e4b-41c5-940d-5768013fdc8d service nova] Acquiring lock "49b7e6c3-9106-4b57-ae44-85b5d4474089-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.735749] env[65121]: DEBUG oslo_concurrency.lockutils [req-24b6397f-9b0f-49f2-bacf-2227b1135902 req-6ee00a30-1e4b-41c5-940d-5768013fdc8d service nova] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.735848] env[65121]: DEBUG oslo_concurrency.lockutils [req-24b6397f-9b0f-49f2-bacf-2227b1135902 req-6ee00a30-1e4b-41c5-940d-5768013fdc8d service nova] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.736015] env[65121]: DEBUG nova.compute.manager [req-24b6397f-9b0f-49f2-bacf-2227b1135902 req-6ee00a30-1e4b-41c5-940d-5768013fdc8d service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] No waiting events found dispatching network-vif-plugged-b2c8d776-fc6a-4893-a904-f33386dc0317 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 920.736195] env[65121]: WARNING nova.compute.manager [req-24b6397f-9b0f-49f2-bacf-2227b1135902 req-6ee00a30-1e4b-41c5-940d-5768013fdc8d service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Received unexpected event network-vif-plugged-b2c8d776-fc6a-4893-a904-f33386dc0317 for instance with vm_state building and task_state spawning. [ 920.761887] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 920.761887] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5217c1fc-c6a4-3533-d158-9aacd6dc6473" [ 920.761887] env[65121]: _type = "HttpNfcLease" [ 920.761887] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 920.768608] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 920.768608] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5217c1fc-c6a4-3533-d158-9aacd6dc6473" [ 920.768608] env[65121]: _type = "HttpNfcLease" [ 920.768608] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 920.769714] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2f0c3d-e12d-402a-9fa2-94ff731524ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.780887] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5228cbcd-28ac-f287-40e2-e322c4cb00fc/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 920.780887] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5228cbcd-28ac-f287-40e2-e322c4cb00fc/disk-0.vmdk. {{(pid=65121) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 920.838410] env[65121]: WARNING neutronclient.v2_0.client [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.839483] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.839934] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.863235] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-71fa5e6e-9c8c-4c06-a790-45cb62259a5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.865224] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c43a90-9a79-8e29-21c9-685f658b1685, 'name': SearchDatastore_Task, 'duration_secs': 0.011833} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.867570] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd3be6bc-a848-4699-9c18-2a31ca403ee8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.875068] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 920.875068] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5288b55d-d694-bd9d-ff20-789787c3c8f0" [ 920.875068] env[65121]: _type = "Task" [ 920.875068] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.886087] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5288b55d-d694-bd9d-ff20-789787c3c8f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.941680] env[65121]: DEBUG nova.network.neutron [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Updated VIF entry in instance network info cache for port 265542f8-48b6-48e7-b0c5-2d371a74f6da. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 920.942123] env[65121]: DEBUG nova.network.neutron [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Updating instance_info_cache with network_info: [{"id": "265542f8-48b6-48e7-b0c5-2d371a74f6da", "address": "fa:16:3e:c3:6b:49", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap265542f8-48", "ovs_interfaceid": "265542f8-48b6-48e7-b0c5-2d371a74f6da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 920.976794] env[65121]: WARNING neutronclient.v2_0.client [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.004718] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106753, 'name': PowerOffVM_Task, 'duration_secs': 0.134275} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.005005] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.005728] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.006818] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3474f491-e3d8-4b9e-a961-7c187fbbf6d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.012000] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.012206] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.012405] env[65121]: DEBUG nova.network.neutron [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 921.017431] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.017637] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0844e76f-4ac9-470b-93ac-944fc7f58731 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.054045] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.054045] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.054045] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Deleting the datastore file [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.054512] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-811aa1d4-d941-4f3d-9dfb-eb31dbb5f6d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.063094] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 921.063094] env[65121]: value = "task-5106756" [ 921.063094] env[65121]: _type = "Task" [ 921.063094] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.073551] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.125785] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "refresh_cache-49b7e6c3-9106-4b57-ae44-85b5d4474089" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.125785] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "refresh_cache-49b7e6c3-9106-4b57-ae44-85b5d4474089" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.126059] env[65121]: DEBUG nova.network.neutron [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 921.200117] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.790s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.200793] env[65121]: DEBUG nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 921.206482] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.144s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.206482] env[65121]: DEBUG nova.objects.instance [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'resources' on Instance uuid cc002961-d742-4255-88c7-f0a5a39424b1 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.220536] env[65121]: DEBUG oslo_vmware.api [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106754, 'name': PowerOffVM_Task, 'duration_secs': 0.166768} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.223242] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.223607] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.223916] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afd4784d-a30d-4f3e-9e34-046b458e10b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.255661] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.257183] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.259313] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.259551] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.259946] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleting the datastore file [datastore1] 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.261874] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac327424-266a-4b9c-be38-6e3e6d7ce3e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.273587] env[65121]: DEBUG oslo_vmware.api [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 921.273587] env[65121]: value = "task-5106758" [ 921.273587] env[65121]: _type = "Task" [ 921.273587] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.288475] env[65121]: DEBUG oslo_vmware.api [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.388823] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5288b55d-d694-bd9d-ff20-789787c3c8f0, 'name': SearchDatastore_Task, 'duration_secs': 0.011015} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.392084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.392374] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c52a1269-bb34-4ef6-ab8d-78df3e9b1d39/c52a1269-bb34-4ef6-ab8d-78df3e9b1d39.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.392684] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55668e9a-9041-4ce5-aa31-5e2b0f56c8a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.406027] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 921.406027] env[65121]: value = "task-5106759" [ 921.406027] env[65121]: _type = "Task" [ 921.406027] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.415313] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106759, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.444618] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Releasing lock "refresh_cache-b2ebce1d-a8d8-4a71-997b-39d09ca373fe" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.444911] env[65121]: DEBUG nova.compute.manager [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Received event network-vif-plugged-98acae83-de55-4870-8650-eb502346b7e4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 921.445202] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Acquiring lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.445627] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.445878] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.446083] env[65121]: DEBUG nova.compute.manager [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] No waiting events found dispatching network-vif-plugged-98acae83-de55-4870-8650-eb502346b7e4 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 921.446276] env[65121]: WARNING nova.compute.manager [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Received unexpected event network-vif-plugged-98acae83-de55-4870-8650-eb502346b7e4 for instance with vm_state building and task_state spawning. [ 921.446457] env[65121]: DEBUG nova.compute.manager [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Received event network-changed-98acae83-de55-4870-8650-eb502346b7e4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 921.446682] env[65121]: DEBUG nova.compute.manager [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Refreshing instance network info cache due to event network-changed-98acae83-de55-4870-8650-eb502346b7e4. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 921.446916] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Acquiring lock "refresh_cache-c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.447029] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Acquired lock "refresh_cache-c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.447208] env[65121]: DEBUG nova.network.neutron [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Refreshing network info cache for port 98acae83-de55-4870-8650-eb502346b7e4 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 921.519477] env[65121]: WARNING neutronclient.v2_0.client [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.519896] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.520174] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.537219] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "600d5554-f52a-48ca-941d-1a755d086823" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.537573] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.537858] env[65121]: DEBUG nova.compute.manager [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 921.538766] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a91a8d-e55d-48da-8bed-2c0012698f97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.548171] env[65121]: DEBUG nova.compute.manager [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 921.549283] env[65121]: DEBUG nova.objects.instance [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'flavor' on Instance uuid 600d5554-f52a-48ca-941d-1a755d086823 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.579533] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106756, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103027} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.582016] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.582259] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.582455] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.631819] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.632312] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.685175] env[65121]: DEBUG nova.network.neutron [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 921.694968] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.694968] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.709156] env[65121]: DEBUG nova.compute.utils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 921.713234] env[65121]: DEBUG nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 921.713478] env[65121]: DEBUG nova.network.neutron [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 921.713816] env[65121]: WARNING neutronclient.v2_0.client [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.714121] env[65121]: WARNING neutronclient.v2_0.client [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.714731] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.715061] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.724598] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.724969] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.761479] env[65121]: DEBUG nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 921.787261] env[65121]: DEBUG oslo_vmware.api [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106758, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.119793} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.791293] env[65121]: DEBUG nova.policy [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a1eec92b3ab4989b6e568736c7a096d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5425f8e2087e4212bdacff24ae2cc0a5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 921.798411] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.798645] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.798832] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.799033] env[65121]: INFO nova.compute.manager [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Took 1.13 seconds to destroy the instance on the hypervisor. [ 921.799315] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 921.802736] env[65121]: DEBUG nova.compute.manager [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 921.802736] env[65121]: DEBUG nova.network.neutron [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 921.802957] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.803609] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.803762] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.837500] env[65121]: WARNING neutronclient.v2_0.client [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.837936] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.838426] env[65121]: WARNING openstack [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.850087] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Completed reading data from the image iterator. {{(pid=65121) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 921.850380] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5228cbcd-28ac-f287-40e2-e322c4cb00fc/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 921.851861] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faad0596-330a-4ba4-9bdf-0a7ab5050f04 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.869213] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5228cbcd-28ac-f287-40e2-e322c4cb00fc/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 921.869383] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5228cbcd-28ac-f287-40e2-e322c4cb00fc/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 921.869652] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ab3748a4-a0b0-44c7-aebf-3c32eb974d18 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.890577] env[65121]: WARNING neutronclient.v2_0.client [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.892018] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.892161] env[65121]: WARNING openstack [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.904493] env[65121]: DEBUG nova.network.neutron [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 921.904628] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.919210] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106759, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50481} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.920097] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] c52a1269-bb34-4ef6-ab8d-78df3e9b1d39/c52a1269-bb34-4ef6-ab8d-78df3e9b1d39.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.921709] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.923632] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15b97fbc-7b8b-4461-b337-39c5a842b019 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.936038] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 921.936038] env[65121]: value = "task-5106760" [ 921.936038] env[65121]: _type = "Task" [ 921.936038] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.950850] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106760, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.955355] env[65121]: WARNING neutronclient.v2_0.client [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.957159] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.959033] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.013676] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 922.014065] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993470', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'name': 'volume-1d017003-f486-4c1b-aa9a-dfd8d2526776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'df5abcb5-583f-4b28-a074-3a3221d74d87', 'attached_at': '', 'detached_at': '', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'serial': '1d017003-f486-4c1b-aa9a-dfd8d2526776'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 922.015130] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9cc19ab-4e5d-418e-b98d-ed7e074c225c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.044104] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f14e98-a763-41d0-aa44-92741837a905 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.075550] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] volume-1d017003-f486-4c1b-aa9a-dfd8d2526776/volume-1d017003-f486-4c1b-aa9a-dfd8d2526776.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.082088] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3e04d63-eee1-4a84-8342-0151814bc3bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.098460] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.101779] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bf150b8-d275-4e58-b68f-1c0c21ebf22e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.110510] env[65121]: DEBUG oslo_vmware.api [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 922.110510] env[65121]: value = "task-5106762" [ 922.110510] env[65121]: _type = "Task" [ 922.110510] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.115823] env[65121]: DEBUG oslo_vmware.api [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 922.115823] env[65121]: value = "task-5106761" [ 922.115823] env[65121]: _type = "Task" [ 922.115823] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.118548] env[65121]: DEBUG nova.network.neutron [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 922.146156] env[65121]: DEBUG oslo_vmware.api [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106761, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.154552] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5228cbcd-28ac-f287-40e2-e322c4cb00fc/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 922.154866] env[65121]: INFO nova.virt.vmwareapi.images [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Downloaded image file data e193384d-79f8-40cc-874e-60556f28628f [ 922.155869] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8250cb1-f050-4dbf-b875-affdb74457b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.177600] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb4997bd-82ca-48f6-b2eb-71deaa8fddf1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.182012] env[65121]: DEBUG nova.network.neutron [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Updating instance_info_cache with network_info: [{"id": "b2c8d776-fc6a-4893-a904-f33386dc0317", "address": "fa:16:3e:09:0a:4d", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2c8d776-fc", "ovs_interfaceid": "b2c8d776-fc6a-4893-a904-f33386dc0317", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 922.206047] env[65121]: INFO nova.virt.vmwareapi.images [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] The imported VM was unregistered [ 922.208970] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Caching image {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 922.208970] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.211116] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef1b1a67-5a9c-4566-a266-047e7a82c056 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.229420] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created directory with path [datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.229775] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38/OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38.vmdk to [datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f/e193384d-79f8-40cc-874e-60556f28628f.vmdk. {{(pid=65121) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 922.230168] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-8edec296-978c-41db-800a-28bd2e615f7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.233620] env[65121]: DEBUG nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 922.243196] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 922.243196] env[65121]: value = "task-5106764" [ 922.243196] env[65121]: _type = "Task" [ 922.243196] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.256366] env[65121]: DEBUG nova.network.neutron [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Successfully created port: afec16fe-3481-4c2e-b503-28fd3f27b085 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 922.261886] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106764, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.280564] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.280969] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.289806] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.413275] env[65121]: DEBUG nova.network.neutron [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 922.434387] env[65121]: WARNING neutronclient.v2_0.client [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 922.434985] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.435381] env[65121]: WARNING openstack [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.459025] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106760, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078131} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.459025] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.459025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca96e09-f0e7-4f01-adbd-ffd485777096 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.482453] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] c52a1269-bb34-4ef6-ab8d-78df3e9b1d39/c52a1269-bb34-4ef6-ab8d-78df3e9b1d39.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.486341] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e02afa0c-ab29-48e9-b313-da29b84b3a7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.513237] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 922.513237] env[65121]: value = "task-5106765" [ 922.513237] env[65121]: _type = "Task" [ 922.513237] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.523234] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10dbd254-7047-4b31-89c3-f7067a7894e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.535030] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1231dc-1136-4da7-bec1-df7efb83b682 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.539121] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106765, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.572434] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9b1d6f-f923-4caf-9092-d5aff08351cc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.583014] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de81449e-2a9f-49d1-b040-453cf7407bfa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.600523] env[65121]: DEBUG nova.compute.provider_tree [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.606021] env[65121]: DEBUG nova.network.neutron [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Updated VIF entry in instance network info cache for port 98acae83-de55-4870-8650-eb502346b7e4. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 922.606021] env[65121]: DEBUG nova.network.neutron [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Updating instance_info_cache with network_info: [{"id": "98acae83-de55-4870-8650-eb502346b7e4", "address": "fa:16:3e:84:29:ab", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98acae83-de", "ovs_interfaceid": "98acae83-de55-4870-8650-eb502346b7e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 922.628496] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.639290] env[65121]: DEBUG oslo_vmware.api [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106762, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.646874] env[65121]: DEBUG oslo_vmware.api [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106761, 'name': PowerOffVM_Task, 'duration_secs': 0.239778} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.649252] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 922.649472] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 922.652396] env[65121]: DEBUG nova.virt.hardware [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 922.652396] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.652396] env[65121]: DEBUG nova.compute.manager [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 922.652396] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3790899d-2589-437f-94be-412c1866c3a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.655269] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24d94aa-1cbe-499c-b965-471e2c52b17a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.667854] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d3aa27-db4e-43e0-9470-4c1d8181dcdb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.688595] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.693293] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 922.693889] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "refresh_cache-49b7e6c3-9106-4b57-ae44-85b5d4474089" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.694238] env[65121]: DEBUG nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Instance network_info: |[{"id": "b2c8d776-fc6a-4893-a904-f33386dc0317", "address": "fa:16:3e:09:0a:4d", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2c8d776-fc", "ovs_interfaceid": "b2c8d776-fc6a-4893-a904-f33386dc0317", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 922.694543] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.694958] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:0a:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2c8d776-fc6a-4893-a904-f33386dc0317', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.702699] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 922.702924] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50ca3118-3a71-44fe-ba20-68eae23e0766 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.716864] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.717607] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ab0d654-152f-4db4-903e-24139ea28862 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.747115] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.747115] env[65121]: value = "task-5106766" [ 922.747115] env[65121]: _type = "Task" [ 922.747115] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.748099] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.748099] env[65121]: value = "task-5106767" [ 922.748099] env[65121]: _type = "Task" [ 922.748099] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.765231] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106766, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.772504] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106767, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.772993] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106764, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.916805] env[65121]: INFO nova.compute.manager [-] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Took 1.11 seconds to deallocate network for instance. [ 922.992339] env[65121]: DEBUG nova.compute.manager [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Received event network-changed-b2c8d776-fc6a-4893-a904-f33386dc0317 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 922.992339] env[65121]: DEBUG nova.compute.manager [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Refreshing instance network info cache due to event network-changed-b2c8d776-fc6a-4893-a904-f33386dc0317. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 922.992339] env[65121]: DEBUG oslo_concurrency.lockutils [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Acquiring lock "refresh_cache-49b7e6c3-9106-4b57-ae44-85b5d4474089" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.992339] env[65121]: DEBUG oslo_concurrency.lockutils [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Acquired lock "refresh_cache-49b7e6c3-9106-4b57-ae44-85b5d4474089" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.992339] env[65121]: DEBUG nova.network.neutron [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Refreshing network info cache for port b2c8d776-fc6a-4893-a904-f33386dc0317 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 923.026836] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106765, 'name': ReconfigVM_Task, 'duration_secs': 0.337587} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.027129] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Reconfigured VM instance instance-00000049 to attach disk [datastore2] c52a1269-bb34-4ef6-ab8d-78df3e9b1d39/c52a1269-bb34-4ef6-ab8d-78df3e9b1d39.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.028140] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0cf1d16-55ed-42fb-8da5-fbc79f8d844a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.039210] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 923.039210] env[65121]: value = "task-5106768" [ 923.039210] env[65121]: _type = "Task" [ 923.039210] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.053448] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106768, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.112025] env[65121]: DEBUG nova.scheduler.client.report [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 923.114790] env[65121]: DEBUG oslo_concurrency.lockutils [req-520645a2-9417-4ea5-9c07-42828b203b41 req-72b15466-3449-442d-8cd8-6b24f09e7b65 service nova] Releasing lock "refresh_cache-c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 923.130763] env[65121]: DEBUG oslo_vmware.api [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106762, 'name': ReconfigVM_Task, 'duration_secs': 0.535576} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.131758] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Reconfigured VM instance instance-00000045 to attach disk [datastore2] volume-1d017003-f486-4c1b-aa9a-dfd8d2526776/volume-1d017003-f486-4c1b-aa9a-dfd8d2526776.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.137456] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06723a29-76c4-4fc3-b242-715494c68cae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.163654] env[65121]: DEBUG oslo_vmware.api [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 923.163654] env[65121]: value = "task-5106769" [ 923.163654] env[65121]: _type = "Task" [ 923.163654] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.178462] env[65121]: DEBUG oslo_vmware.api [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106769, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.179124] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12f58717-4f23-4b09-8f1e-3f09377e98a5 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.641s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.188978] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34fa43f2-5ae1-4ff9-99f9-26a7cf8c61fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.217015] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583ad169-e6ac-43f5-a71a-f801bdc44eed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.232158] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance '7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 923.249348] env[65121]: DEBUG nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 923.276387] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106767, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.283959] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106764, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.284656] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106766, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.295794] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 923.296142] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 923.296307] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 923.296516] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 923.296832] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 923.297058] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 923.297289] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.297441] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 923.297613] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 923.297906] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 923.298166] env[65121]: DEBUG nova.virt.hardware [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 923.302864] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b26273-2afb-4bde-a2c9-cedaead365bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.318920] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5439505d-0f0d-44bf-a225-56cd89cf4335 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.428937] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 923.495125] env[65121]: WARNING neutronclient.v2_0.client [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.495852] env[65121]: WARNING openstack [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.496232] env[65121]: WARNING openstack [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.553860] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106768, 'name': Rename_Task, 'duration_secs': 0.190618} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.554257] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.554443] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f586615-710b-4da1-a0c9-6e4a3c5b6a93 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.566790] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 923.566790] env[65121]: value = "task-5106770" [ 923.566790] env[65121]: _type = "Task" [ 923.566790] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.582734] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.618103] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.412s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.620020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.251s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 923.621856] env[65121]: INFO nova.compute.claims [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.644934] env[65121]: INFO nova.scheduler.client.report [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted allocations for instance cc002961-d742-4255-88c7-f0a5a39424b1 [ 923.680296] env[65121]: DEBUG oslo_vmware.api [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106769, 'name': ReconfigVM_Task, 'duration_secs': 0.34184} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.680723] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993470', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'name': 'volume-1d017003-f486-4c1b-aa9a-dfd8d2526776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'df5abcb5-583f-4b28-a074-3a3221d74d87', 'attached_at': '', 'detached_at': '', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'serial': '1d017003-f486-4c1b-aa9a-dfd8d2526776'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 923.724643] env[65121]: INFO nova.compute.manager [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Rebuilding instance [ 923.739305] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c8fc0fd-f409-43be-a460-99f544db3a91 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance '7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 923.746560] env[65121]: WARNING openstack [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.746560] env[65121]: WARNING openstack [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.780852] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106764, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.783329] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106766, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.788263] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106767, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.795649] env[65121]: DEBUG nova.compute.manager [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 923.796650] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69706a6f-b136-4cc4-8356-cd5e9d4949cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.851866] env[65121]: WARNING neutronclient.v2_0.client [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.852721] env[65121]: WARNING openstack [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.853124] env[65121]: WARNING openstack [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.929257] env[65121]: DEBUG nova.network.neutron [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Successfully updated port: afec16fe-3481-4c2e-b503-28fd3f27b085 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 923.993409] env[65121]: DEBUG nova.network.neutron [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Updated VIF entry in instance network info cache for port b2c8d776-fc6a-4893-a904-f33386dc0317. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 923.993949] env[65121]: DEBUG nova.network.neutron [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Updating instance_info_cache with network_info: [{"id": "b2c8d776-fc6a-4893-a904-f33386dc0317", "address": "fa:16:3e:09:0a:4d", "network": {"id": "3e8b08a8-0b4b-4a74-9d13-357ced3a64b2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-679796700-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b35af4e2b6844a9081f34466661c83b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2c8d776-fc", "ovs_interfaceid": "b2c8d776-fc6a-4893-a904-f33386dc0317", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 924.082027] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106770, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.153730] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ad8d3395-4574-4982-9822-5e05d1b3850c tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "cc002961-d742-4255-88c7-f0a5a39424b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.443s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 924.272352] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106764, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.285127] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106766, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.285127] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106767, 'name': CreateVM_Task, 'duration_secs': 1.177538} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.285287] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.285825] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.285999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 924.286397] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 924.286705] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b890a6cc-f6ea-49bd-9fb2-09ea76572a1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.296506] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 924.296506] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52148548-8e74-807f-224f-698f88d00d77" [ 924.296506] env[65121]: _type = "Task" [ 924.296506] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.317843] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52148548-8e74-807f-224f-698f88d00d77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.434577] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "refresh_cache-ac001ee6-f601-48f7-af0d-42c2387f3c98" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.435047] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquired lock "refresh_cache-ac001ee6-f601-48f7-af0d-42c2387f3c98" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 924.435288] env[65121]: DEBUG nova.network.neutron [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 924.499922] env[65121]: DEBUG oslo_concurrency.lockutils [req-669a8f15-61ae-44c8-bfa2-3072446ae216 req-6096f7f3-68b8-4ba8-b34b-19fe6cfe3af0 service nova] Releasing lock "refresh_cache-49b7e6c3-9106-4b57-ae44-85b5d4474089" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 924.581456] env[65121]: DEBUG oslo_vmware.api [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106770, 'name': PowerOnVM_Task, 'duration_secs': 0.55842} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.581755] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.581978] env[65121]: INFO nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Took 7.88 seconds to spawn the instance on the hypervisor. [ 924.582204] env[65121]: DEBUG nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 924.583088] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77ddb3a-cf8e-422c-a444-4ab171d8baef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.726805] env[65121]: DEBUG nova.objects.instance [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'flavor' on Instance uuid df5abcb5-583f-4b28-a074-3a3221d74d87 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.771120] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106764, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.778544] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106766, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.813511] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52148548-8e74-807f-224f-698f88d00d77, 'name': SearchDatastore_Task, 'duration_secs': 0.046326} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.813690] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 924.814267] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.814267] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.814366] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 924.814448] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.814728] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a14fcf2-31db-412c-9980-856de951a10f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.817766] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.817766] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a0c1df4-19d9-4306-ab3c-1cabb6cb81a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.838861] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 924.838861] env[65121]: value = "task-5106771" [ 924.838861] env[65121]: _type = "Task" [ 924.838861] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.838861] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.838861] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.840582] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff43c01f-e24e-44d6-af2a-c23e1cf52746 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.856117] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 924.856423] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.856804] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 924.856804] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524a66b2-d366-dc10-c7eb-9cacf742d1b5" [ 924.856804] env[65121]: _type = "Task" [ 924.856804] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.857997] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd4176f-fc57-448b-ac12-0fa181360781 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.874023] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.876787] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c4b4055-3bdb-43aa-b2fc-104f22738605 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.879552] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524a66b2-d366-dc10-c7eb-9cacf742d1b5, 'name': SearchDatastore_Task, 'duration_secs': 0.014127} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.883975] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe5be9be-5fb0-434d-9817-5f526e706fd9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.891857] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 924.891857] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522d2459-e642-1e2e-2d4d-2622092623e9" [ 924.891857] env[65121]: _type = "Task" [ 924.891857] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.903782] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522d2459-e642-1e2e-2d4d-2622092623e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.945560] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.945989] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.976055] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.976055] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.976055] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleting the datastore file [datastore1] 600d5554-f52a-48ca-941d-1a755d086823 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.980815] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a9034dd-eb5c-43f9-81ce-bbd11c799ebe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.991065] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "df5abcb5-583f-4b28-a074-3a3221d74d87" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.001258] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 925.001258] env[65121]: value = "task-5106773" [ 925.001258] env[65121]: _type = "Task" [ 925.001258] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.010564] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.038892] env[65121]: DEBUG nova.network.neutron [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 925.088072] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.088462] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.107590] env[65121]: INFO nova.compute.manager [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Took 44.45 seconds to build instance. [ 925.214781] env[65121]: WARNING neutronclient.v2_0.client [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.215748] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.216060] env[65121]: WARNING openstack [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.232720] env[65121]: DEBUG oslo_concurrency.lockutils [None req-81fa0fae-ea67-4bd3-ad6e-441f5c4c39f1 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.884s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.233781] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.243s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.234016] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "df5abcb5-583f-4b28-a074-3a3221d74d87-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.234231] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.234959] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.238877] env[65121]: INFO nova.compute.manager [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Terminating instance [ 925.263488] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106764, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.691718} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.266762] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38/OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38.vmdk to [datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f/e193384d-79f8-40cc-874e-60556f28628f.vmdk. [ 925.266979] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Cleaning up location [datastore2] OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 925.267160] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_71fac137-9f8c-4b66-947a-dbe3a0760a38 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.271183] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-328b27c5-1297-44c8-9bf4-ea518e82d8cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.285621] env[65121]: DEBUG nova.compute.manager [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Received event network-vif-plugged-afec16fe-3481-4c2e-b503-28fd3f27b085 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 925.286197] env[65121]: DEBUG oslo_concurrency.lockutils [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Acquiring lock "ac001ee6-f601-48f7-af0d-42c2387f3c98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.286433] env[65121]: DEBUG oslo_concurrency.lockutils [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.286833] env[65121]: DEBUG oslo_concurrency.lockutils [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.286833] env[65121]: DEBUG nova.compute.manager [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] No waiting events found dispatching network-vif-plugged-afec16fe-3481-4c2e-b503-28fd3f27b085 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 925.286912] env[65121]: WARNING nova.compute.manager [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Received unexpected event network-vif-plugged-afec16fe-3481-4c2e-b503-28fd3f27b085 for instance with vm_state building and task_state spawning. [ 925.287064] env[65121]: DEBUG nova.compute.manager [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Received event network-changed-afec16fe-3481-4c2e-b503-28fd3f27b085 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 925.287236] env[65121]: DEBUG nova.compute.manager [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Refreshing instance network info cache due to event network-changed-afec16fe-3481-4c2e-b503-28fd3f27b085. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 925.287368] env[65121]: DEBUG oslo_concurrency.lockutils [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Acquiring lock "refresh_cache-ac001ee6-f601-48f7-af0d-42c2387f3c98" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.296279] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106766, 'name': CreateVM_Task, 'duration_secs': 2.213991} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.296663] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 925.296663] env[65121]: value = "task-5106774" [ 925.296663] env[65121]: _type = "Task" [ 925.296663] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.303176] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.304258] env[65121]: WARNING neutronclient.v2_0.client [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.304639] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.304782] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.305115] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 925.309893] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5b18c6d-b9a4-430a-adfd-ae5129d6ee24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.313114] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8163f7fc-1fc9-4b33-a25f-b18ff3803be1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.322365] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.326137] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 925.326137] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52603715-e3f4-2fa1-ccdf-4f4796407557" [ 925.326137] env[65121]: _type = "Task" [ 925.326137] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.327153] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4521daf2-b1ae-48f6-81b5-667ee7fe0bc3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.341537] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52603715-e3f4-2fa1-ccdf-4f4796407557, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.368495] env[65121]: DEBUG nova.network.neutron [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Updating instance_info_cache with network_info: [{"id": "afec16fe-3481-4c2e-b503-28fd3f27b085", "address": "fa:16:3e:83:dc:bb", "network": {"id": "1dcca16a-e688-4489-9b1f-01bbdc25fdfc", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1639899199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5425f8e2087e4212bdacff24ae2cc0a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4b6ddb2-2e19-4031-9b22-add90d41a114", "external-id": "nsx-vlan-transportzone-921", "segmentation_id": 921, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapafec16fe-34", "ovs_interfaceid": "afec16fe-3481-4c2e-b503-28fd3f27b085", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 925.370326] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910977df-c75b-42ad-9120-6cde1b827461 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.381077] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc30b411-e0aa-4afa-ba61-09996dc8487d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.399739] env[65121]: DEBUG nova.compute.provider_tree [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.410331] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522d2459-e642-1e2e-2d4d-2622092623e9, 'name': SearchDatastore_Task, 'duration_secs': 0.013327} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.410589] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.410842] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 925.411171] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65d1d4e3-d72c-4bcd-9416-fe331d9268f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.420369] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 925.420369] env[65121]: value = "task-5106775" [ 925.420369] env[65121]: _type = "Task" [ 925.420369] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.431937] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.512912] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.49351} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.513214] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.513390] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.513650] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.609984] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4dae7b26-7f97-4617-9a14-1d12b7d0fdb4 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.965s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.743874] env[65121]: DEBUG nova.compute.manager [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 925.744191] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.744577] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58660aae-405a-40b2-88d7-767fd55f8565 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.753236] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 925.753236] env[65121]: value = "task-5106776" [ 925.753236] env[65121]: _type = "Task" [ 925.753236] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.762047] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.814583] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14435} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.814857] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.815032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f/e193384d-79f8-40cc-874e-60556f28628f.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.815288] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f/e193384d-79f8-40cc-874e-60556f28628f.vmdk to [datastore2] b2ebce1d-a8d8-4a71-997b-39d09ca373fe/b2ebce1d-a8d8-4a71-997b-39d09ca373fe.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 925.815688] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1debe9ac-3a85-4f83-b645-176d1a7e92a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.824917] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 925.824917] env[65121]: value = "task-5106777" [ 925.824917] env[65121]: _type = "Task" [ 925.824917] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.835232] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.849578] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52603715-e3f4-2fa1-ccdf-4f4796407557, 'name': SearchDatastore_Task, 'duration_secs': 0.035803} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.849914] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.850156] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.850398] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.850538] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.850760] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.851101] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31b7d890-88db-48ad-b1ec-0858c05c0fc4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.867331] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.867510] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.868836] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73bb0047-b59a-4002-bcf0-9ece31dcfd53 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.877918] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Releasing lock "refresh_cache-ac001ee6-f601-48f7-af0d-42c2387f3c98" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.877918] env[65121]: DEBUG nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Instance network_info: |[{"id": "afec16fe-3481-4c2e-b503-28fd3f27b085", "address": "fa:16:3e:83:dc:bb", "network": {"id": "1dcca16a-e688-4489-9b1f-01bbdc25fdfc", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1639899199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5425f8e2087e4212bdacff24ae2cc0a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4b6ddb2-2e19-4031-9b22-add90d41a114", "external-id": "nsx-vlan-transportzone-921", "segmentation_id": 921, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapafec16fe-34", "ovs_interfaceid": "afec16fe-3481-4c2e-b503-28fd3f27b085", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 925.877918] env[65121]: DEBUG oslo_concurrency.lockutils [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Acquired lock "refresh_cache-ac001ee6-f601-48f7-af0d-42c2387f3c98" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.877918] env[65121]: DEBUG nova.network.neutron [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Refreshing network info cache for port afec16fe-3481-4c2e-b503-28fd3f27b085 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 925.879013] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:dc:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4b6ddb2-2e19-4031-9b22-add90d41a114', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'afec16fe-3481-4c2e-b503-28fd3f27b085', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.886720] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Creating folder: Project (5425f8e2087e4212bdacff24ae2cc0a5). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.887096] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 925.887096] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52246d80-202f-a35e-337d-8a4d5495b310" [ 925.887096] env[65121]: _type = "Task" [ 925.887096] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.888203] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e2920ee-befa-409b-bb6d-bd72081f3b3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.901193] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52246d80-202f-a35e-337d-8a4d5495b310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.906360] env[65121]: DEBUG nova.scheduler.client.report [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 925.912101] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Created folder: Project (5425f8e2087e4212bdacff24ae2cc0a5) in parent group-v993268. [ 925.912417] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Creating folder: Instances. Parent ref: group-v993476. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 925.913719] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bcb08cc-c006-47cb-a2ad-7b7105f24184 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.933716] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Created folder: Instances in parent group-v993476. [ 925.934064] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 925.934595] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.934825] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b602d6c-b205-41d3-8937-b0a3a0335a41 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.956039] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.957701] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.957970] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.958193] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.958415] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.958522] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.962015] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.962015] env[65121]: value = "task-5106780" [ 925.962015] env[65121]: _type = "Task" [ 925.962015] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.966143] env[65121]: INFO nova.compute.manager [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Terminating instance [ 925.975887] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106780, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.266252] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106776, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.336691] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.389629] env[65121]: WARNING neutronclient.v2_0.client [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 926.390345] env[65121]: WARNING openstack [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.390706] env[65121]: WARNING openstack [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.414910] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.795s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.415285] env[65121]: DEBUG nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 926.419492] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52246d80-202f-a35e-337d-8a4d5495b310, 'name': SearchDatastore_Task, 'duration_secs': 0.052703} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.419492] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.284s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.419492] env[65121]: DEBUG nova.objects.instance [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lazy-loading 'resources' on Instance uuid 93721179-407a-43d9-b0bf-157433a09519 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.421715] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06d47d89-8eb6-4cb2-a9f2-7d0c1d75e7ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.428494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.428838] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.428960] env[65121]: DEBUG nova.compute.manager [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Going to confirm migration 3 {{(pid=65121) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 926.436504] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 926.436504] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524f14f9-1bb8-ec90-cc6a-ca210bc69026" [ 926.436504] env[65121]: _type = "Task" [ 926.436504] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.441108] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106775, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.452464] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524f14f9-1bb8-ec90-cc6a-ca210bc69026, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.479320] env[65121]: DEBUG nova.compute.manager [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 926.479534] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.480956] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106780, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.481762] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd99ded-ec4a-4fd5-a1c1-c0791f1ddbb4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.496863] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.497545] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27b5f532-27e5-4f89-b744-bb29ba8f8bde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.507486] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 926.507486] env[65121]: value = "task-5106781" [ 926.507486] env[65121]: _type = "Task" [ 926.507486] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.516475] env[65121]: WARNING openstack [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.516894] env[65121]: WARNING openstack [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.536033] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.574789] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 926.575174] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 926.575343] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 926.575583] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 926.575756] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 926.575890] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 926.578684] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.578684] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 926.578684] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 926.578684] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 926.578684] env[65121]: DEBUG nova.virt.hardware [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 926.578684] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4fe386-77d9-4618-befa-ca6839a8a966 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.594898] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f0a087-dc06-444e-8e98-b0a49360012c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.617466] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:20:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2c37a16-3107-4510-9c9d-17bf130945ff', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.626120] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 926.627978] env[65121]: WARNING neutronclient.v2_0.client [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 926.629297] env[65121]: WARNING openstack [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.629297] env[65121]: WARNING openstack [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.638627] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.639279] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71aaa188-73f1-4506-95c3-3a7afffaf301 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.661980] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.661980] env[65121]: value = "task-5106782" [ 926.661980] env[65121]: _type = "Task" [ 926.661980] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.672139] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106782, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.738958] env[65121]: DEBUG nova.network.neutron [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Updated VIF entry in instance network info cache for port afec16fe-3481-4c2e-b503-28fd3f27b085. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 926.739808] env[65121]: DEBUG nova.network.neutron [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Updating instance_info_cache with network_info: [{"id": "afec16fe-3481-4c2e-b503-28fd3f27b085", "address": "fa:16:3e:83:dc:bb", "network": {"id": "1dcca16a-e688-4489-9b1f-01bbdc25fdfc", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1639899199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5425f8e2087e4212bdacff24ae2cc0a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4b6ddb2-2e19-4031-9b22-add90d41a114", "external-id": "nsx-vlan-transportzone-921", "segmentation_id": 921, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapafec16fe-34", "ovs_interfaceid": "afec16fe-3481-4c2e-b503-28fd3f27b085", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 926.768584] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106776, 'name': PowerOffVM_Task, 'duration_secs': 0.560833} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.768797] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.768973] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 926.769210] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993470', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'name': 'volume-1d017003-f486-4c1b-aa9a-dfd8d2526776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'df5abcb5-583f-4b28-a074-3a3221d74d87', 'attached_at': '', 'detached_at': '', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'serial': '1d017003-f486-4c1b-aa9a-dfd8d2526776'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 926.770379] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b206e26-66ef-4d37-b3a6-0a3dba5dcea6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.800483] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a96d2e-f801-4184-958f-e2e9b1df5d4a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.810174] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967ed8e5-5d1d-429f-bbdd-b97686f4c59b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.839592] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088a1c3c-056f-42a4-8a18-bc22cd1281a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.849023] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.862789] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The volume has not been displaced from its original location: [datastore2] volume-1d017003-f486-4c1b-aa9a-dfd8d2526776/volume-1d017003-f486-4c1b-aa9a-dfd8d2526776.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 926.868870] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 926.869300] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdb2b72c-943b-4d3a-9280-ad5aaa45ac7d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.892218] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 926.892218] env[65121]: value = "task-5106783" [ 926.892218] env[65121]: _type = "Task" [ 926.892218] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.905538] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106783, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.927406] env[65121]: DEBUG nova.compute.utils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 926.933079] env[65121]: DEBUG nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 926.933079] env[65121]: DEBUG nova.network.neutron [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 926.933079] env[65121]: WARNING neutronclient.v2_0.client [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 926.933079] env[65121]: WARNING neutronclient.v2_0.client [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 926.933079] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.933962] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.954309] env[65121]: WARNING neutronclient.v2_0.client [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 926.975893] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524f14f9-1bb8-ec90-cc6a-ca210bc69026, 'name': SearchDatastore_Task, 'duration_secs': 0.037482} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.975893] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106775, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.1772} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.975893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.976157] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 49b7e6c3-9106-4b57-ae44-85b5d4474089/49b7e6c3-9106-4b57-ae44-85b5d4474089.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.976456] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.976755] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.978367] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b78cd81f-310b-40f7-ab37-0c6bb1f0eb14 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.980087] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce6d4b0c-e8ab-4592-a6f8-0da7b75f6e2d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.989698] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106780, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.997100] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 926.997100] env[65121]: value = "task-5106784" [ 926.997100] env[65121]: _type = "Task" [ 926.997100] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.999204] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 926.999204] env[65121]: value = "task-5106785" [ 926.999204] env[65121]: _type = "Task" [ 926.999204] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.017188] env[65121]: DEBUG nova.policy [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '175d752afbc745258500be8624bfbf14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e491b88dcaad42dcb345b3217bc6f941', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 927.030279] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.035987] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106785, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.042803] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106781, 'name': PowerOffVM_Task, 'duration_secs': 0.237939} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.043324] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.043484] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.043770] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eacdc9c5-698e-48a5-87f6-15398d39fed6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.053611] env[65121]: WARNING neutronclient.v2_0.client [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.053611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.053611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.053611] env[65121]: DEBUG nova.network.neutron [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 927.054140] env[65121]: DEBUG nova.objects.instance [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'info_cache' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.134356] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.134630] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.134911] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleting the datastore file [datastore2] c52a1269-bb34-4ef6-ab8d-78df3e9b1d39 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.135341] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05d4f473-8589-4247-a245-9934d8df1e9c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.149296] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 927.149296] env[65121]: value = "task-5106787" [ 927.149296] env[65121]: _type = "Task" [ 927.149296] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.163039] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.179133] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106782, 'name': CreateVM_Task, 'duration_secs': 0.441723} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.179322] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.180763] env[65121]: WARNING neutronclient.v2_0.client [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 927.180763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.180763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.180918] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 927.181122] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-909fad75-262f-4889-a272-5d6cbe255907 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.190252] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 927.190252] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52be1fc0-92b7-c03d-a038-fd6a299ce05f" [ 927.190252] env[65121]: _type = "Task" [ 927.190252] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.202128] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.202452] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "23d6f632-8918-46af-b239-08a9615dfbec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.215957] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be1fc0-92b7-c03d-a038-fd6a299ce05f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.243119] env[65121]: DEBUG oslo_concurrency.lockutils [req-b368f8ea-aa80-43c8-9fa6-13e4645b3c59 req-2383e375-7e20-4151-bcfa-4beb153b89fe service nova] Releasing lock "refresh_cache-ac001ee6-f601-48f7-af0d-42c2387f3c98" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 927.364606] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.404251] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106783, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.444687] env[65121]: DEBUG nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 927.479579] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106780, 'name': CreateVM_Task, 'duration_secs': 1.276599} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.479813] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.480833] env[65121]: WARNING neutronclient.v2_0.client [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.480833] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.482287] env[65121]: DEBUG nova.network.neutron [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Successfully created port: 920563c9-c663-4389-9348-0d9ccbca5fea {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 927.519545] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106785, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107506} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.520152] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.520406] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.521463] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72a1281-7292-4547-ae8c-1ec92dd325aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.552681] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.553751] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c92c28e8-d062-4f81-a7e1-63361d43a147 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.585035] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 927.585035] env[65121]: value = "task-5106788" [ 927.585035] env[65121]: _type = "Task" [ 927.585035] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.596628] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106788, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.648868] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c473b9ea-ab78-42e9-8500-21287227d1e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.662659] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.665873] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a87ba04-1168-4100-a9d0-9b4532fbfd03 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.707897] env[65121]: DEBUG nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 927.714616] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c676b08-9076-4671-af13-a0abf72f1e69 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.729644] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9f91cc-1082-4303-8435-6003fd13d303 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.734100] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be1fc0-92b7-c03d-a038-fd6a299ce05f, 'name': SearchDatastore_Task, 'duration_secs': 0.060218} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.735086] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 927.735086] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.735371] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.735423] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.735589] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.736341] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.737227] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 927.737227] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45e28d47-55ea-4b8b-ab29-6f6aac077189 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.739354] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64374f61-3dcd-4570-87dc-318788c2cf19 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.751128] env[65121]: DEBUG nova.compute.provider_tree [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.759781] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 927.759781] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521b045a-3719-f0e4-fe5e-5435d9879289" [ 927.759781] env[65121]: _type = "Task" [ 927.759781] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.765785] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.766086] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.767999] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5bc64d2-f516-46f9-817a-e474bb072ce9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.775190] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521b045a-3719-f0e4-fe5e-5435d9879289, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.779921] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 927.779921] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d15427-6a24-a66d-aa89-d404792e1ebc" [ 927.779921] env[65121]: _type = "Task" [ 927.779921] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.791512] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d15427-6a24-a66d-aa89-d404792e1ebc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.848550] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.906396] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106783, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.011781] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.076855] env[65121]: WARNING neutronclient.v2_0.client [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.078214] env[65121]: WARNING openstack [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.079211] env[65121]: WARNING openstack [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.099630] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.166430] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.234220] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.254678] env[65121]: DEBUG nova.scheduler.client.report [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 928.279312] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521b045a-3719-f0e4-fe5e-5435d9879289, 'name': SearchDatastore_Task, 'duration_secs': 0.054135} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.281032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.281634] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.281966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.303128] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d15427-6a24-a66d-aa89-d404792e1ebc, 'name': SearchDatastore_Task, 'duration_secs': 0.049292} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.304400] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cf7fa36-e662-41d3-a952-0494d20da9ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.313310] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 928.313310] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529da189-4f01-5279-5d1f-8c66d02cf097" [ 928.313310] env[65121]: _type = "Task" [ 928.313310] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.323935] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529da189-4f01-5279-5d1f-8c66d02cf097, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.348056] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.405967] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106783, 'name': ReconfigVM_Task, 'duration_secs': 1.040707} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.406303] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 928.411998] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-896c9939-80a8-4a03-a260-96620b3bdf18 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.432105] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 928.432105] env[65121]: value = "task-5106789" [ 928.432105] env[65121]: _type = "Task" [ 928.432105] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.442183] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106789, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.456177] env[65121]: DEBUG nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 928.460594] env[65121]: WARNING openstack [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.460975] env[65121]: WARNING openstack [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.487982] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 928.487982] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 928.487982] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 928.487982] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 928.488330] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 928.488330] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 928.488391] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.488548] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 928.488706] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 928.488945] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 928.489143] env[65121]: DEBUG nova.virt.hardware [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 928.490083] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05fd5b3d-e0ff-4613-9551-7d14648ea6ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.501039] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee00658-a3a0-4d53-beb9-b6bea8c0478c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.529138] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.601613] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.636953] env[65121]: WARNING neutronclient.v2_0.client [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.637735] env[65121]: WARNING openstack [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.638073] env[65121]: WARNING openstack [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.666715] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.761839] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.343s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.764530] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.982s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.765149] env[65121]: DEBUG nova.objects.instance [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 928.785235] env[65121]: INFO nova.scheduler.client.report [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted allocations for instance 93721179-407a-43d9-b0bf-157433a09519 [ 928.822129] env[65121]: DEBUG nova.network.neutron [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 928.833483] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529da189-4f01-5279-5d1f-8c66d02cf097, 'name': SearchDatastore_Task, 'duration_secs': 0.048717} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.833483] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.833483] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.833805] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 928.833805] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.834024] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5cacf241-8d04-48c6-a6eb-e6275b6e40ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.836947] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a4dd846-fee0-43ba-bd33-0da0567b9aa2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.857893] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 928.857893] env[65121]: value = "task-5106790" [ 928.857893] env[65121]: _type = "Task" [ 928.857893] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.857893] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.861595] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.861595] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.861595] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e1ccd8d-e70d-4e25-bae4-711e13ae0c0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.874023] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.879443] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 928.879443] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5221f2d1-2a98-cb24-8bce-845249441f9d" [ 928.879443] env[65121]: _type = "Task" [ 928.879443] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.891386] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5221f2d1-2a98-cb24-8bce-845249441f9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.945211] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106789, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.019295] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106784, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.105213] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106788, 'name': ReconfigVM_Task, 'duration_secs': 1.500039} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.106067] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e/2e676b64-181e-4fee-8120-05cb49b5bb4e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.106964] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8a98551-e926-477d-9b43-2d89d8e4e308 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.115939] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 929.115939] env[65121]: value = "task-5106791" [ 929.115939] env[65121]: _type = "Task" [ 929.115939] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.129716] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106791, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.166077] env[65121]: DEBUG oslo_vmware.api [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106787, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.900719} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.166313] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.166504] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.167019] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.167019] env[65121]: INFO nova.compute.manager [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Took 2.69 seconds to destroy the instance on the hypervisor. [ 929.167790] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 929.167790] env[65121]: DEBUG nova.compute.manager [-] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 929.167790] env[65121]: DEBUG nova.network.neutron [-] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 929.168096] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.168497] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.168790] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.294829] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3221c5d-7cb7-4b06-80e2-6c994582b791 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "93721179-407a-43d9-b0bf-157433a09519" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.743s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.297916] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.327244] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.327471] env[65121]: DEBUG nova.objects.instance [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'migration_context' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.342508] env[65121]: DEBUG nova.network.neutron [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Successfully updated port: 920563c9-c663-4389-9348-0d9ccbca5fea {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 929.360665] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106777, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.1605} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.361572] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e193384d-79f8-40cc-874e-60556f28628f/e193384d-79f8-40cc-874e-60556f28628f.vmdk to [datastore2] b2ebce1d-a8d8-4a71-997b-39d09ca373fe/b2ebce1d-a8d8-4a71-997b-39d09ca373fe.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.362815] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09db54ae-4e93-40d7-b38c-6c3813582fb9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.371524] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.398797] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] b2ebce1d-a8d8-4a71-997b-39d09ca373fe/b2ebce1d-a8d8-4a71-997b-39d09ca373fe.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.403324] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47b2d079-326d-414a-84ad-72e366984ad1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.428387] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5221f2d1-2a98-cb24-8bce-845249441f9d, 'name': SearchDatastore_Task, 'duration_secs': 0.056015} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.430720] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 929.430720] env[65121]: value = "task-5106792" [ 929.430720] env[65121]: _type = "Task" [ 929.430720] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.431170] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c30c1db-70d8-4a25-ac07-3be214390a34 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.449243] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106792, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.457028] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106789, 'name': ReconfigVM_Task, 'duration_secs': 0.571777} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.457028] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 929.457028] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525dd2d3-6bda-9f73-83a7-030fcc5bad4f" [ 929.457028] env[65121]: _type = "Task" [ 929.457028] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.457028] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993470', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'name': 'volume-1d017003-f486-4c1b-aa9a-dfd8d2526776', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'df5abcb5-583f-4b28-a074-3a3221d74d87', 'attached_at': '', 'detached_at': '', 'volume_id': '1d017003-f486-4c1b-aa9a-dfd8d2526776', 'serial': '1d017003-f486-4c1b-aa9a-dfd8d2526776'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 929.457028] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.457775] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fc70de-6d62-4fd6-83b7-e4d27821a268 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.472193] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525dd2d3-6bda-9f73-83a7-030fcc5bad4f, 'name': SearchDatastore_Task, 'duration_secs': 0.019418} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.474556] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.474843] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] ac001ee6-f601-48f7-af0d-42c2387f3c98/ac001ee6-f601-48f7-af0d-42c2387f3c98.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.476628] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.476628] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4b0c286-fdf7-47df-93db-087c9e21fc0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.479165] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26113b9a-cc00-416c-a54d-7bcc6d00a08f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.487252] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 929.487252] env[65121]: value = "task-5106793" [ 929.487252] env[65121]: _type = "Task" [ 929.487252] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.498456] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106793, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.518036] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106784, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.458898} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.518228] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 49b7e6c3-9106-4b57-ae44-85b5d4474089/49b7e6c3-9106-4b57-ae44-85b5d4474089.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.518306] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.519057] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb6e7b03-3dc7-4bb9-bddb-2813d3d6fb09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.525809] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 929.525809] env[65121]: value = "task-5106795" [ 929.525809] env[65121]: _type = "Task" [ 929.525809] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.535719] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106795, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.544266] env[65121]: DEBUG nova.compute.manager [req-11a364f7-5efb-40fc-9e48-0837305d0fde req-c3eee066-84ea-4a56-a9be-8193a88eeb24 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Received event network-vif-plugged-920563c9-c663-4389-9348-0d9ccbca5fea {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 929.544731] env[65121]: DEBUG oslo_concurrency.lockutils [req-11a364f7-5efb-40fc-9e48-0837305d0fde req-c3eee066-84ea-4a56-a9be-8193a88eeb24 service nova] Acquiring lock "59d6005d-6caf-4898-b791-70d9c015cdb4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 929.545274] env[65121]: DEBUG oslo_concurrency.lockutils [req-11a364f7-5efb-40fc-9e48-0837305d0fde req-c3eee066-84ea-4a56-a9be-8193a88eeb24 service nova] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.546278] env[65121]: DEBUG oslo_concurrency.lockutils [req-11a364f7-5efb-40fc-9e48-0837305d0fde req-c3eee066-84ea-4a56-a9be-8193a88eeb24 service nova] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.546278] env[65121]: DEBUG nova.compute.manager [req-11a364f7-5efb-40fc-9e48-0837305d0fde req-c3eee066-84ea-4a56-a9be-8193a88eeb24 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] No waiting events found dispatching network-vif-plugged-920563c9-c663-4389-9348-0d9ccbca5fea {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 929.546469] env[65121]: WARNING nova.compute.manager [req-11a364f7-5efb-40fc-9e48-0837305d0fde req-c3eee066-84ea-4a56-a9be-8193a88eeb24 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Received unexpected event network-vif-plugged-920563c9-c663-4389-9348-0d9ccbca5fea for instance with vm_state building and task_state spawning. [ 929.558484] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.558910] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.559257] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleting the datastore file [datastore1] df5abcb5-583f-4b28-a074-3a3221d74d87 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.559710] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7bd7e595-7dde-4da5-a882-4f787ef61aed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.571953] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 929.571953] env[65121]: value = "task-5106796" [ 929.571953] env[65121]: _type = "Task" [ 929.571953] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.587189] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.630058] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106791, 'name': Rename_Task, 'duration_secs': 0.434995} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.630411] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.630708] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-406c8bc8-997e-45d0-97e0-b44a69f6fd99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.642349] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 929.642349] env[65121]: value = "task-5106797" [ 929.642349] env[65121]: _type = "Task" [ 929.642349] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.659040] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106797, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.769851] env[65121]: DEBUG nova.compute.manager [req-1a729e3f-c2e6-479e-acd4-9384ca07407e req-a22400c8-75e1-4c02-9288-017c8596f271 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Received event network-vif-deleted-98acae83-de55-4870-8650-eb502346b7e4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 929.770277] env[65121]: INFO nova.compute.manager [req-1a729e3f-c2e6-479e-acd4-9384ca07407e req-a22400c8-75e1-4c02-9288-017c8596f271 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Neutron deleted interface 98acae83-de55-4870-8650-eb502346b7e4; detaching it from the instance and deleting it from the info cache [ 929.770520] env[65121]: DEBUG nova.network.neutron [req-1a729e3f-c2e6-479e-acd4-9384ca07407e req-a22400c8-75e1-4c02-9288-017c8596f271 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 929.776690] env[65121]: DEBUG oslo_concurrency.lockutils [None req-159e658d-d3e1-409f-aaa6-d9c28f2f7cf2 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.778621] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.554s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.778819] env[65121]: DEBUG nova.objects.instance [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lazy-loading 'resources' on Instance uuid 8a23da27-ce6c-4453-9036-65eeeda3ce0d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.830988] env[65121]: DEBUG nova.objects.base [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Object Instance<7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c> lazy-loaded attributes: info_cache,migration_context {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 929.832183] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a9a01a-5bad-4496-ab67-f7ec71f31f67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.855060] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-59d6005d-6caf-4898-b791-70d9c015cdb4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.855299] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-59d6005d-6caf-4898-b791-70d9c015cdb4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.855511] env[65121]: DEBUG nova.network.neutron [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 929.861472] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bbb4305-b466-49fa-93f4-4fc94a649c76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.870878] env[65121]: DEBUG oslo_vmware.api [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 929.870878] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527a9f3d-1b3d-9a6b-7717-4feb264056cc" [ 929.870878] env[65121]: _type = "Task" [ 929.870878] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.875284] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106790, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.886760] env[65121]: DEBUG oslo_vmware.api [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527a9f3d-1b3d-9a6b-7717-4feb264056cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010666} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.887103] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 929.948375] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106792, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.998496] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106793, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.044547] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106795, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.28272} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.044547] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.045124] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ece3b9-b839-4c47-8830-adb506658ca6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.074908] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 49b7e6c3-9106-4b57-ae44-85b5d4474089/49b7e6c3-9106-4b57-ae44-85b5d4474089.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.075345] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0adc3984-095c-4eb7-9093-687db1bac23a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.101206] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.102960] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 930.102960] env[65121]: value = "task-5106798" [ 930.102960] env[65121]: _type = "Task" [ 930.102960] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.116563] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106798, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.142434] env[65121]: DEBUG nova.network.neutron [-] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 930.154495] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106797, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.275033] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08ad76c9-c597-4aa4-b508-fd6b9fd22df1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.287629] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac60479-7c9c-489f-a503-36809eed8822 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.332910] env[65121]: DEBUG nova.compute.manager [req-1a729e3f-c2e6-479e-acd4-9384ca07407e req-a22400c8-75e1-4c02-9288-017c8596f271 service nova] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Detach interface failed, port_id=98acae83-de55-4870-8650-eb502346b7e4, reason: Instance c52a1269-bb34-4ef6-ab8d-78df3e9b1d39 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 930.370033] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.370460] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.377509] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106790, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.028374} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.380158] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.380377] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.381101] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-448a567e-284a-4121-a690-ef68e2717255 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.389018] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 930.389018] env[65121]: value = "task-5106799" [ 930.389018] env[65121]: _type = "Task" [ 930.389018] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.400650] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106799, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.421513] env[65121]: DEBUG nova.network.neutron [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 930.440499] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.441047] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.453746] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106792, 'name': ReconfigVM_Task, 'duration_secs': 0.53079} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.453917] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Reconfigured VM instance instance-00000048 to attach disk [datastore2] b2ebce1d-a8d8-4a71-997b-39d09ca373fe/b2ebce1d-a8d8-4a71-997b-39d09ca373fe.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.454597] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-41a8e7d8-054a-4007-95e8-cb1184e00129 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.462736] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 930.462736] env[65121]: value = "task-5106800" [ 930.462736] env[65121]: _type = "Task" [ 930.462736] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.476176] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106800, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.501871] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106793, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.831724} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.502384] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] ac001ee6-f601-48f7-af0d-42c2387f3c98/ac001ee6-f601-48f7-af0d-42c2387f3c98.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.502800] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.503277] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-16770a8f-a94e-40b0-864b-cabf02faf0e9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.513041] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 930.513041] env[65121]: value = "task-5106801" [ 930.513041] env[65121]: _type = "Task" [ 930.513041] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.523993] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106801, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.528990] env[65121]: WARNING neutronclient.v2_0.client [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.529835] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.530356] env[65121]: WARNING openstack [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.586462] env[65121]: DEBUG oslo_vmware.api [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106796, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.562597} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.591435] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.591640] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.591817] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.591984] env[65121]: INFO nova.compute.manager [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Took 4.85 seconds to destroy the instance on the hypervisor. [ 930.592251] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 930.595179] env[65121]: DEBUG nova.compute.manager [-] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 930.595288] env[65121]: DEBUG nova.network.neutron [-] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 930.595528] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.596098] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.596358] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.615773] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106798, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.636445] env[65121]: DEBUG nova.network.neutron [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Updating instance_info_cache with network_info: [{"id": "920563c9-c663-4389-9348-0d9ccbca5fea", "address": "fa:16:3e:11:23:f8", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920563c9-c6", "ovs_interfaceid": "920563c9-c663-4389-9348-0d9ccbca5fea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 930.651508] env[65121]: INFO nova.compute.manager [-] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Took 1.48 seconds to deallocate network for instance. [ 930.663137] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.663727] env[65121]: DEBUG oslo_vmware.api [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106797, 'name': PowerOnVM_Task, 'duration_secs': 0.935746} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.666586] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 930.666719] env[65121]: DEBUG nova.compute.manager [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 930.668673] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc4e35e-9c04-4f05-bc83-28e5fa45b556 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.834757] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e321aa6-49b6-4d13-b6b6-70cb4be67bca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.845641] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688328c7-2ae5-4dc8-9ba8-1018fc12169d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.879177] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0befaf-c47e-4c1f-b14a-8d8eea7c7f28 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.889611] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e82616-e0bc-4def-9f79-48f68534067b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.909824] env[65121]: DEBUG nova.compute.provider_tree [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.915100] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106799, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078878} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.915795] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.916687] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1dddb9-abe5-4dba-967a-9d9e74a3b14e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.943373] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.944330] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92ce00c1-ff2b-45f5-b14a-da4de0e18d7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.968441] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 930.968441] env[65121]: value = "task-5106802" [ 930.968441] env[65121]: _type = "Task" [ 930.968441] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.978378] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106800, 'name': Rename_Task, 'duration_secs': 0.159519} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.979067] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.979324] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7148b69f-6120-4c76-964b-01cab84fb54b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.984238] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106802, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.989441] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 930.989441] env[65121]: value = "task-5106803" [ 930.989441] env[65121]: _type = "Task" [ 930.989441] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.002405] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106803, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.023236] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106801, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182248} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.024274] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.024405] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d073fa-4acd-49c5-9311-56bb0aac7bc5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.050443] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] ac001ee6-f601-48f7-af0d-42c2387f3c98/ac001ee6-f601-48f7-af0d-42c2387f3c98.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.051733] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a313d41-973f-4a6d-9a61-f2326cf1dbe3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.072342] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 931.072342] env[65121]: value = "task-5106804" [ 931.072342] env[65121]: _type = "Task" [ 931.072342] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.081684] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106804, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.115913] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106798, 'name': ReconfigVM_Task, 'duration_secs': 0.771626} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.115913] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 49b7e6c3-9106-4b57-ae44-85b5d4474089/49b7e6c3-9106-4b57-ae44-85b5d4474089.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.116215] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96ffdd24-3630-4e0f-844a-b68053f5b229 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.124267] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 931.124267] env[65121]: value = "task-5106805" [ 931.124267] env[65121]: _type = "Task" [ 931.124267] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.141428] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-59d6005d-6caf-4898-b791-70d9c015cdb4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.141703] env[65121]: DEBUG nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Instance network_info: |[{"id": "920563c9-c663-4389-9348-0d9ccbca5fea", "address": "fa:16:3e:11:23:f8", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920563c9-c6", "ovs_interfaceid": "920563c9-c663-4389-9348-0d9ccbca5fea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 931.142053] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106805, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.142421] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:23:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '920563c9-c663-4389-9348-0d9ccbca5fea', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.150583] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 931.150840] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.151082] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-287692fe-7de6-4ea4-a11f-ddc49dc42de0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.169346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.176114] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.176114] env[65121]: value = "task-5106806" [ 931.176114] env[65121]: _type = "Task" [ 931.176114] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.188037] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106806, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.190094] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.384682] env[65121]: DEBUG nova.network.neutron [-] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 931.416767] env[65121]: DEBUG nova.scheduler.client.report [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 931.481470] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106802, 'name': ReconfigVM_Task, 'duration_secs': 0.357939} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.481798] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 600d5554-f52a-48ca-941d-1a755d086823/600d5554-f52a-48ca-941d-1a755d086823.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.482722] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7ea4fb1-7692-4c03-b9d7-b6409f4b05d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.492113] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 931.492113] env[65121]: value = "task-5106807" [ 931.492113] env[65121]: _type = "Task" [ 931.492113] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.510705] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106803, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.511099] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106807, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.584440] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106804, 'name': ReconfigVM_Task, 'duration_secs': 0.488484} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.584769] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Reconfigured VM instance instance-0000004b to attach disk [datastore1] ac001ee6-f601-48f7-af0d-42c2387f3c98/ac001ee6-f601-48f7-af0d-42c2387f3c98.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.585443] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-54745cdd-ac5f-4fa6-8834-42c7a7e3a47b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.594702] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 931.594702] env[65121]: value = "task-5106808" [ 931.594702] env[65121]: _type = "Task" [ 931.594702] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.606551] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106808, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.635196] env[65121]: DEBUG nova.compute.manager [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Received event network-changed-920563c9-c663-4389-9348-0d9ccbca5fea {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 931.635443] env[65121]: DEBUG nova.compute.manager [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Refreshing instance network info cache due to event network-changed-920563c9-c663-4389-9348-0d9ccbca5fea. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 931.635714] env[65121]: DEBUG oslo_concurrency.lockutils [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Acquiring lock "refresh_cache-59d6005d-6caf-4898-b791-70d9c015cdb4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.635881] env[65121]: DEBUG oslo_concurrency.lockutils [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Acquired lock "refresh_cache-59d6005d-6caf-4898-b791-70d9c015cdb4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.636071] env[65121]: DEBUG nova.network.neutron [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Refreshing network info cache for port 920563c9-c663-4389-9348-0d9ccbca5fea {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 931.641894] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106805, 'name': Rename_Task, 'duration_secs': 0.216865} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.642553] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.642871] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b60b8c32-f22e-4665-b3fb-8296067e0d03 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.651036] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 931.651036] env[65121]: value = "task-5106809" [ 931.651036] env[65121]: _type = "Task" [ 931.651036] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.662659] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106809, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.689322] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106806, 'name': CreateVM_Task, 'duration_secs': 0.497395} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.689533] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.690146] env[65121]: WARNING neutronclient.v2_0.client [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 931.690550] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.690731] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.691075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 931.691350] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6248915f-ff66-44a3-92b8-ba9c1ac60cef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.697985] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 931.697985] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ff6a6a-bdc7-a440-5794-9434945ccd6d" [ 931.697985] env[65121]: _type = "Task" [ 931.697985] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.708390] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ff6a6a-bdc7-a440-5794-9434945ccd6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.860875] env[65121]: DEBUG nova.compute.manager [req-5bbaecb6-a220-4d25-bcdc-46f17e025d12 req-f172c998-5342-4cd9-8568-8107aaf6cbb8 service nova] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Received event network-vif-deleted-9e8f421a-1606-4c5a-821d-90bee69ab953 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 931.888072] env[65121]: INFO nova.compute.manager [-] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Took 1.29 seconds to deallocate network for instance. [ 931.923623] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.145s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.926966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.570s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.927397] env[65121]: DEBUG nova.objects.instance [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lazy-loading 'resources' on Instance uuid 99b1ff9c-38b7-4947-81bb-d200a3b2a09a {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.956159] env[65121]: INFO nova.scheduler.client.report [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Deleted allocations for instance 8a23da27-ce6c-4453-9036-65eeeda3ce0d [ 932.008814] env[65121]: DEBUG oslo_vmware.api [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106803, 'name': PowerOnVM_Task, 'duration_secs': 0.53532} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.013100] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.013382] env[65121]: INFO nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Took 18.16 seconds to spawn the instance on the hypervisor. [ 932.013604] env[65121]: DEBUG nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 932.013921] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106807, 'name': Rename_Task, 'duration_secs': 0.208131} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.015169] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2418dee3-687f-405c-9dc2-22547b3abe95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.017696] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.017997] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1c93b7e-6894-4db8-8086-85afb2604718 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.031577] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 932.031577] env[65121]: value = "task-5106810" [ 932.031577] env[65121]: _type = "Task" [ 932.031577] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.044930] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106810, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.106725] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106808, 'name': Rename_Task, 'duration_secs': 0.215329} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.107251] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.108711] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ad30d2e-df64-444c-8f93-2bf968d5525a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.111859] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "2e676b64-181e-4fee-8120-05cb49b5bb4e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.112105] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "2e676b64-181e-4fee-8120-05cb49b5bb4e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.112346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "2e676b64-181e-4fee-8120-05cb49b5bb4e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.112475] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "2e676b64-181e-4fee-8120-05cb49b5bb4e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.112633] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "2e676b64-181e-4fee-8120-05cb49b5bb4e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.116350] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 932.116350] env[65121]: value = "task-5106811" [ 932.116350] env[65121]: _type = "Task" [ 932.116350] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.117116] env[65121]: INFO nova.compute.manager [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Terminating instance [ 932.130984] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106811, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.144224] env[65121]: WARNING neutronclient.v2_0.client [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.144612] env[65121]: WARNING openstack [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.145032] env[65121]: WARNING openstack [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.166110] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106809, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.210917] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ff6a6a-bdc7-a440-5794-9434945ccd6d, 'name': SearchDatastore_Task, 'duration_secs': 0.010973} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.211243] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 932.211471] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.211778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.211929] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.212135] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.212750] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2df85be0-aaea-47f4-84ea-a28156ecc1eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.222570] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.222813] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.226518] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbef132f-984e-4851-b395-c0a16d8ef5a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.233337] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 932.233337] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5234caee-acf0-ceb0-c934-437c75b0219a" [ 932.233337] env[65121]: _type = "Task" [ 932.233337] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.243972] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5234caee-acf0-ceb0-c934-437c75b0219a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.280673] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "7fcab9cb-9a0f-4a50-9244-687889443a31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.281104] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.281781] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "7fcab9cb-9a0f-4a50-9244-687889443a31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.281781] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.281781] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.290520] env[65121]: INFO nova.compute.manager [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Terminating instance [ 932.313810] env[65121]: WARNING openstack [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.314468] env[65121]: WARNING openstack [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.395171] env[65121]: WARNING neutronclient.v2_0.client [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.396060] env[65121]: WARNING openstack [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.396508] env[65121]: WARNING openstack [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.464833] env[65121]: DEBUG oslo_concurrency.lockutils [None req-513c3083-726f-433f-86d9-ea56182eb56a tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "8a23da27-ce6c-4453-9036-65eeeda3ce0d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.654s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.478594] env[65121]: INFO nova.compute.manager [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Took 0.59 seconds to detach 1 volumes for instance. [ 932.547549] env[65121]: DEBUG nova.network.neutron [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Updated VIF entry in instance network info cache for port 920563c9-c663-4389-9348-0d9ccbca5fea. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 932.547549] env[65121]: DEBUG nova.network.neutron [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Updating instance_info_cache with network_info: [{"id": "920563c9-c663-4389-9348-0d9ccbca5fea", "address": "fa:16:3e:11:23:f8", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap920563c9-c6", "ovs_interfaceid": "920563c9-c663-4389-9348-0d9ccbca5fea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 932.553726] env[65121]: INFO nova.compute.manager [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Took 58.35 seconds to build instance. [ 932.564633] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106810, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.625888] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "refresh_cache-2e676b64-181e-4fee-8120-05cb49b5bb4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.626087] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquired lock "refresh_cache-2e676b64-181e-4fee-8120-05cb49b5bb4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.626264] env[65121]: DEBUG nova.network.neutron [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 932.631439] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106811, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.666097] env[65121]: DEBUG oslo_vmware.api [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106809, 'name': PowerOnVM_Task, 'duration_secs': 0.764609} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.668655] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.668875] env[65121]: INFO nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Took 12.75 seconds to spawn the instance on the hypervisor. [ 932.669072] env[65121]: DEBUG nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 932.670240] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b1aea3-ca43-4134-9c3d-2556077cbf8f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.747945] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5234caee-acf0-ceb0-c934-437c75b0219a, 'name': SearchDatastore_Task, 'duration_secs': 0.013845} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.748780] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28b59b95-0e43-4cc2-a97c-6f0fdabe82cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.757264] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 932.757264] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5202ac13-2a66-62a9-4418-91c5705376c6" [ 932.757264] env[65121]: _type = "Task" [ 932.757264] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.766715] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5202ac13-2a66-62a9-4418-91c5705376c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.796478] env[65121]: DEBUG nova.compute.manager [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 932.796683] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.797777] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adbd6c5-c054-46bf-994b-50acd74735b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.808055] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.810831] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fb5f23e-4a8a-4f6e-88dc-a87448576156 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.818973] env[65121]: DEBUG oslo_vmware.api [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 932.818973] env[65121]: value = "task-5106812" [ 932.818973] env[65121]: _type = "Task" [ 932.818973] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.831223] env[65121]: DEBUG oslo_vmware.api [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.836197] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.965682] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e563ea72-d56c-444f-a985-5689d9fb9749 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.974814] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c382bf7-d761-48a8-8e1e-add9df1dc6f2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.011630] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.013561] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60757d5e-9c26-4b17-bb37-d2dd37182190 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.026857] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a36a069-7dab-4baf-8071-8b361c2a61ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.048403] env[65121]: DEBUG nova.compute.provider_tree [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.054222] env[65121]: DEBUG oslo_concurrency.lockutils [req-7def975a-f2ae-45b0-b195-3cd7ba2f6679 req-9e0bdd92-828d-42d0-936c-223f856fa9f0 service nova] Releasing lock "refresh_cache-59d6005d-6caf-4898-b791-70d9c015cdb4" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.060749] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7d8bba0-2517-46a6-8e41-7ae4aab1b401 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.611s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.061943] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106810, 'name': PowerOnVM_Task, 'duration_secs': 0.559526} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.062210] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.226s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.062430] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.062630] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.062798] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.068089] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.068387] env[65121]: DEBUG nova.compute.manager [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 933.069181] env[65121]: INFO nova.compute.manager [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Terminating instance [ 933.071984] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba167efc-ff27-40a7-b631-f4462cb83dcd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.129792] env[65121]: WARNING neutronclient.v2_0.client [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.130805] env[65121]: WARNING openstack [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 933.131250] env[65121]: WARNING openstack [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.145835] env[65121]: DEBUG oslo_vmware.api [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106811, 'name': PowerOnVM_Task, 'duration_secs': 0.620762} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.147533] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.147533] env[65121]: INFO nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Took 9.90 seconds to spawn the instance on the hypervisor. [ 933.147533] env[65121]: DEBUG nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 933.147533] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88d6d24-96ae-412f-9c3f-0538262f0c8f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.163287] env[65121]: DEBUG nova.network.neutron [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 933.193532] env[65121]: INFO nova.compute.manager [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Took 47.73 seconds to build instance. [ 933.238504] env[65121]: DEBUG nova.network.neutron [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 933.269601] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5202ac13-2a66-62a9-4418-91c5705376c6, 'name': SearchDatastore_Task, 'duration_secs': 0.02553} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.269934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.270225] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.270517] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba62a756-9e00-48a2-a7fe-8ddab1f95565 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.278864] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 933.278864] env[65121]: value = "task-5106813" [ 933.278864] env[65121]: _type = "Task" [ 933.278864] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.289949] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.330826] env[65121]: DEBUG oslo_vmware.api [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106812, 'name': PowerOffVM_Task, 'duration_secs': 0.466861} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.331292] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.331292] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.331549] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce67eb7d-9716-4dd1-b3f6-d34a6fe1050a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.413025] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.413610] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.413887] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Deleting the datastore file [datastore1] 7fcab9cb-9a0f-4a50-9244-687889443a31 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.414323] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-519699b6-5ff5-4005-ac36-61eed8b470e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.425364] env[65121]: DEBUG oslo_vmware.api [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for the task: (returnval){ [ 933.425364] env[65121]: value = "task-5106815" [ 933.425364] env[65121]: _type = "Task" [ 933.425364] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.435591] env[65121]: DEBUG oslo_vmware.api [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.519927] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.520226] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.520349] env[65121]: INFO nova.compute.manager [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Shelving [ 933.555179] env[65121]: DEBUG nova.scheduler.client.report [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 933.581354] env[65121]: DEBUG nova.compute.manager [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 933.581354] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.581354] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1ab209-26ae-488a-939d-ecadb74f8c56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.587547] env[65121]: INFO nova.compute.manager [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] bringing vm to original state: 'stopped' [ 933.593219] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.593219] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b436005f-ddb4-41cc-ba22-403e77eeb13d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.604300] env[65121]: DEBUG oslo_vmware.api [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 933.604300] env[65121]: value = "task-5106816" [ 933.604300] env[65121]: _type = "Task" [ 933.604300] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.617745] env[65121]: DEBUG oslo_vmware.api [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106816, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.673146] env[65121]: INFO nova.compute.manager [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Took 44.10 seconds to build instance. [ 933.697176] env[65121]: DEBUG oslo_concurrency.lockutils [None req-40a76dab-edc0-4f02-857b-9ef6382c41f5 tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.236s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.742171] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Releasing lock "refresh_cache-2e676b64-181e-4fee-8120-05cb49b5bb4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.742755] env[65121]: DEBUG nova.compute.manager [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 933.742957] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.744098] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2a4e4f-df8e-432a-ad11-2866e6dcfd3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.757432] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.757705] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e627d0a-9aa1-4eeb-8d24-7a296d69e551 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.769311] env[65121]: DEBUG oslo_vmware.api [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 933.769311] env[65121]: value = "task-5106817" [ 933.769311] env[65121]: _type = "Task" [ 933.769311] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.780518] env[65121]: DEBUG oslo_vmware.api [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.792334] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106813, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.804073] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.804073] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.804073] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.804073] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.804513] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.807579] env[65121]: INFO nova.compute.manager [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Terminating instance [ 933.937765] env[65121]: DEBUG oslo_vmware.api [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Task: {'id': task-5106815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.499312} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.938109] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.938304] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.938502] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.938683] env[65121]: INFO nova.compute.manager [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Took 1.14 seconds to destroy the instance on the hypervisor. [ 933.938943] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 933.939170] env[65121]: DEBUG nova.compute.manager [-] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 933.939260] env[65121]: DEBUG nova.network.neutron [-] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 933.939519] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.940144] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 933.940460] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.996129] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.064027] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.137s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.066613] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.197s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.068606] env[65121]: INFO nova.compute.claims [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.097421] env[65121]: INFO nova.scheduler.client.report [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted allocations for instance 99b1ff9c-38b7-4947-81bb-d200a3b2a09a [ 934.118071] env[65121]: DEBUG oslo_vmware.api [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106816, 'name': PowerOffVM_Task, 'duration_secs': 0.251813} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.118401] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.118557] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.118814] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8466f660-577a-4c4c-813d-930290f5252c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.126944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "49b7e6c3-9106-4b57-ae44-85b5d4474089" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.127223] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.127424] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "49b7e6c3-9106-4b57-ae44-85b5d4474089-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.127683] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.127934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.132808] env[65121]: INFO nova.compute.manager [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Terminating instance [ 934.175217] env[65121]: DEBUG oslo_concurrency.lockutils [None req-333774b5-d205-4aa3-b605-21d9263175ce tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.609s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.280598] env[65121]: DEBUG oslo_vmware.api [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106817, 'name': PowerOffVM_Task, 'duration_secs': 0.277569} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.285327] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.285538] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.285897] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e54d2a9-c504-4918-893d-2a49a4f22429 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.294812] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627129} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.294812] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.294922] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.296505] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-938adbaf-e90b-4c21-b58c-85d85a53cfc8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.304108] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 934.304108] env[65121]: value = "task-5106820" [ 934.304108] env[65121]: _type = "Task" [ 934.304108] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.316319] env[65121]: DEBUG nova.compute.manager [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 934.317220] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.317354] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106820, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.320314] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e25641-b287-449c-8a3f-45efe3b6b904 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.322826] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.323065] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.323285] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Deleting the datastore file [datastore1] 2e676b64-181e-4fee-8120-05cb49b5bb4e {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.323653] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58c13cae-0ae9-461b-9540-b5c85a4b80e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.333747] env[65121]: DEBUG oslo_vmware.api [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for the task: (returnval){ [ 934.333747] env[65121]: value = "task-5106821" [ 934.333747] env[65121]: _type = "Task" [ 934.333747] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.336260] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.339754] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec0bec88-1ea3-41e6-8393-4ac81af21af0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.347845] env[65121]: DEBUG oslo_vmware.api [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106821, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.352033] env[65121]: DEBUG oslo_vmware.api [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 934.352033] env[65121]: value = "task-5106822" [ 934.352033] env[65121]: _type = "Task" [ 934.352033] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.358586] env[65121]: DEBUG oslo_vmware.api [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.534795] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.534795] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c94e81fb-6df3-4033-bc28-484fb6e75fcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.543210] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 934.543210] env[65121]: value = "task-5106823" [ 934.543210] env[65121]: _type = "Task" [ 934.543210] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.563686] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106823, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.599922] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "600d5554-f52a-48ca-941d-1a755d086823" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.600198] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.600378] env[65121]: DEBUG nova.compute.manager [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 934.604728] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6ad614-ab8d-4db1-acdd-2eb2a9db296b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.608874] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ad88269-b045-47b1-9f2c-cb82e4033f14 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "99b1ff9c-38b7-4947-81bb-d200a3b2a09a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.664s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.614482] env[65121]: DEBUG nova.compute.manager [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 934.638772] env[65121]: DEBUG nova.compute.manager [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 934.638772] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.638772] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933b87eb-ca01-4443-b0c1-22907c314062 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.648710] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.649055] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1993d3f-9845-476e-9b20-1a44db1d8f79 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.658592] env[65121]: DEBUG oslo_vmware.api [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 934.658592] env[65121]: value = "task-5106824" [ 934.658592] env[65121]: _type = "Task" [ 934.658592] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.669560] env[65121]: DEBUG oslo_vmware.api [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.745917] env[65121]: DEBUG nova.compute.manager [req-f0a57f6b-e443-4284-ba7b-8aabf91e15ed req-5b11d0a4-c080-4d93-95df-a4c8c08724cc service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Received event network-vif-deleted-296c4d11-0b03-45a7-8019-6f23e26fea6e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 934.745917] env[65121]: INFO nova.compute.manager [req-f0a57f6b-e443-4284-ba7b-8aabf91e15ed req-5b11d0a4-c080-4d93-95df-a4c8c08724cc service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Neutron deleted interface 296c4d11-0b03-45a7-8019-6f23e26fea6e; detaching it from the instance and deleting it from the info cache [ 934.745917] env[65121]: DEBUG nova.network.neutron [req-f0a57f6b-e443-4284-ba7b-8aabf91e15ed req-5b11d0a4-c080-4d93-95df-a4c8c08724cc service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 934.818577] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106820, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098142} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.818885] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.819829] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c32ed5d-95bc-46be-97fe-d50586282867 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.848796] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.852427] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dea6b6fe-cb9a-4de5-84c5-c82643d5c81e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.880542] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "b6f6871d-311c-4adb-824e-2907a12f4224" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.880732] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "b6f6871d-311c-4adb-824e-2907a12f4224" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.881101] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "b6f6871d-311c-4adb-824e-2907a12f4224-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.881258] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "b6f6871d-311c-4adb-824e-2907a12f4224-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.881476] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "b6f6871d-311c-4adb-824e-2907a12f4224-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.883162] env[65121]: DEBUG oslo_vmware.api [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Task: {'id': task-5106821, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258268} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.887885] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.887885] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.888085] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.888134] env[65121]: INFO nova.compute.manager [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 934.888347] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 934.888959] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 934.888959] env[65121]: value = "task-5106825" [ 934.888959] env[65121]: _type = "Task" [ 934.888959] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.888959] env[65121]: DEBUG oslo_vmware.api [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106822, 'name': PowerOffVM_Task, 'duration_secs': 0.234884} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.889398] env[65121]: INFO nova.compute.manager [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Terminating instance [ 934.890961] env[65121]: DEBUG nova.compute.manager [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 934.891079] env[65121]: DEBUG nova.network.neutron [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 934.891299] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.891851] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.892419] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.899283] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.899514] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.904786] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e25ebd61-ad60-4bf0-80b8-825a200f2386 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.915620] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.931945] env[65121]: DEBUG nova.network.neutron [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 934.932262] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.055399] env[65121]: DEBUG nova.network.neutron [-] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 935.056882] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106823, 'name': PowerOffVM_Task, 'duration_secs': 0.393586} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.057465] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.058329] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd075db-f245-48e2-8b5b-4a0fde4a8fc1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.085860] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54edde76-0037-4d0e-8833-9715170da6d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.120094] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.123281] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8122fa2f-99e9-4fd9-bbd3-c1221c09f390 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.129381] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 935.129381] env[65121]: value = "task-5106827" [ 935.129381] env[65121]: _type = "Task" [ 935.129381] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.142361] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.179415] env[65121]: DEBUG oslo_vmware.api [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106824, 'name': PowerOffVM_Task, 'duration_secs': 0.333312} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.182721] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.182721] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.183816] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47a0aa6e-d670-4240-a8eb-e6c5441e5847 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.250165] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecf2508a-3993-4d48-a455-7a3fe4a14151 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.266458] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cff8f8-2cbd-4eb6-bd7e-a73addc8df55 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.315045] env[65121]: DEBUG nova.compute.manager [req-f0a57f6b-e443-4284-ba7b-8aabf91e15ed req-5b11d0a4-c080-4d93-95df-a4c8c08724cc service nova] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Detach interface failed, port_id=296c4d11-0b03-45a7-8019-6f23e26fea6e, reason: Instance 7fcab9cb-9a0f-4a50-9244-687889443a31 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 935.345631] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.346032] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.346338] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore2] b2ebce1d-a8d8-4a71-997b-39d09ca373fe {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.346950] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb2f4a70-d92c-4246-adec-8b47ae429880 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.357496] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.357832] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.358097] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleting the datastore file [datastore2] 49b7e6c3-9106-4b57-ae44-85b5d4474089 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.358427] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-389ef954-b533-4b2c-bb70-4bfa0001e3e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.364975] env[65121]: DEBUG oslo_vmware.api [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 935.364975] env[65121]: value = "task-5106829" [ 935.364975] env[65121]: _type = "Task" [ 935.364975] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.367225] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.367345] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.367449] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Deleting the datastore file [datastore2] ab41ae41-e69d-47fb-a31a-16748fd3b0f2 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.371470] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75844297-bdeb-44f2-8ae4-359d5136f570 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.378035] env[65121]: DEBUG oslo_vmware.api [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for the task: (returnval){ [ 935.378035] env[65121]: value = "task-5106830" [ 935.378035] env[65121]: _type = "Task" [ 935.378035] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.386305] env[65121]: DEBUG oslo_vmware.api [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106829, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.388437] env[65121]: DEBUG oslo_vmware.api [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for the task: (returnval){ [ 935.388437] env[65121]: value = "task-5106831" [ 935.388437] env[65121]: _type = "Task" [ 935.388437] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.395520] env[65121]: DEBUG oslo_vmware.api [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.406559] env[65121]: DEBUG oslo_vmware.api [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106831, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.407872] env[65121]: DEBUG nova.compute.manager [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 935.408026] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.409139] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9242428-4f66-4948-aac3-0ec5ee20c2e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.417255] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106825, 'name': ReconfigVM_Task, 'duration_secs': 0.531369} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.418033] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.419230] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d13b26b1-6b6e-4dbe-a567-4d20e0cfe8a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.425600] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.426498] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a84521a9-a16e-438b-b1a4-6387380ba0ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.433203] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 935.433203] env[65121]: value = "task-5106832" [ 935.433203] env[65121]: _type = "Task" [ 935.433203] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.437954] env[65121]: DEBUG nova.network.neutron [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 935.439868] env[65121]: DEBUG oslo_vmware.api [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 935.439868] env[65121]: value = "task-5106833" [ 935.439868] env[65121]: _type = "Task" [ 935.439868] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.450636] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106832, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.456808] env[65121]: DEBUG oslo_vmware.api [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.562040] env[65121]: INFO nova.compute.manager [-] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Took 1.62 seconds to deallocate network for instance. [ 935.603157] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 935.605213] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-76c707ec-37fa-4820-81ae-dd181745f949 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.617457] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 935.617457] env[65121]: value = "task-5106834" [ 935.617457] env[65121]: _type = "Task" [ 935.617457] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.638442] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106834, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.647170] env[65121]: DEBUG oslo_vmware.api [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106827, 'name': PowerOffVM_Task, 'duration_secs': 0.245119} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.647480] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.647637] env[65121]: DEBUG nova.compute.manager [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 935.648518] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca42acaf-cc91-439b-9797-a8316023ba3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.726026] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afebfa5e-ee69-40be-bcab-9cae2a2e2782 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.732468] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49207fc-4ee4-4a24-8af6-068e321fff52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.767033] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529a405a-7668-485e-9732-edd2a5a525a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.775717] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16332f7f-b053-4956-a3e8-c66b01c7cb66 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.790512] env[65121]: DEBUG nova.compute.provider_tree [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.878558] env[65121]: DEBUG oslo_vmware.api [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106829, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284795} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.878764] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.878945] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.879195] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.879307] env[65121]: INFO nova.compute.manager [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Took 2.30 seconds to destroy the instance on the hypervisor. [ 935.879590] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 935.879830] env[65121]: DEBUG nova.compute.manager [-] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 935.879970] env[65121]: DEBUG nova.network.neutron [-] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 935.880257] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.880811] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.881087] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.900494] env[65121]: DEBUG oslo_vmware.api [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Task: {'id': task-5106831, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398804} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.903774] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.903990] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.904189] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.904364] env[65121]: INFO nova.compute.manager [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Took 1.59 seconds to destroy the instance on the hypervisor. [ 935.904606] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 935.904831] env[65121]: DEBUG oslo_vmware.api [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Task: {'id': task-5106830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299374} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.905059] env[65121]: DEBUG nova.compute.manager [-] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 935.905159] env[65121]: DEBUG nova.network.neutron [-] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 935.905413] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.906052] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.906323] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.913023] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.914123] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.914123] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.914123] env[65121]: INFO nova.compute.manager [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Took 1.28 seconds to destroy the instance on the hypervisor. [ 935.914123] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 935.914775] env[65121]: DEBUG nova.compute.manager [-] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 935.914775] env[65121]: DEBUG nova.network.neutron [-] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 935.914775] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.916106] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.916106] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.925527] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.940819] env[65121]: INFO nova.compute.manager [-] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Took 1.05 seconds to deallocate network for instance. [ 935.950201] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106832, 'name': Rename_Task, 'duration_secs': 0.184158} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.951140] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.955513] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.956629] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-480e8b92-81cb-4777-9347-9853ec55ae65 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.962170] env[65121]: DEBUG oslo_vmware.api [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106833, 'name': PowerOffVM_Task, 'duration_secs': 0.210298} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.962421] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.962596] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.964120] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77fb2641-7fa9-447d-a211-99778c91df7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.965969] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 935.965969] env[65121]: value = "task-5106835" [ 935.965969] env[65121]: _type = "Task" [ 935.965969] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.971499] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.979564] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106835, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.060036] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.060036] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.061267] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleting the datastore file [datastore2] b6f6871d-311c-4adb-824e-2907a12f4224 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.061267] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2187d9ba-ac57-483a-8997-a34948c6988a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.071967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.071967] env[65121]: DEBUG oslo_vmware.api [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 936.071967] env[65121]: value = "task-5106837" [ 936.071967] env[65121]: _type = "Task" [ 936.071967] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.089108] env[65121]: DEBUG oslo_vmware.api [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106837, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.133330] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106834, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.165130] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.565s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.294418] env[65121]: DEBUG nova.scheduler.client.report [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 936.458638] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.481389] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106835, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.564223] env[65121]: DEBUG nova.compute.manager [req-34fb7fab-9467-4d70-bf94-b2a30ccea515 req-73791ea8-4cfb-4af2-84e1-7abd71555856 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Received event network-vif-deleted-1740aa3b-2428-4bc3-8888-c4edc0734c60 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 936.564325] env[65121]: INFO nova.compute.manager [req-34fb7fab-9467-4d70-bf94-b2a30ccea515 req-73791ea8-4cfb-4af2-84e1-7abd71555856 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Neutron deleted interface 1740aa3b-2428-4bc3-8888-c4edc0734c60; detaching it from the instance and deleting it from the info cache [ 936.564471] env[65121]: DEBUG nova.network.neutron [req-34fb7fab-9467-4d70-bf94-b2a30ccea515 req-73791ea8-4cfb-4af2-84e1-7abd71555856 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 936.584123] env[65121]: DEBUG oslo_vmware.api [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106837, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211575} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.584375] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.584799] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 936.584998] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 936.585221] env[65121]: INFO nova.compute.manager [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Took 1.18 seconds to destroy the instance on the hypervisor. [ 936.585671] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 936.587062] env[65121]: DEBUG nova.compute.manager [-] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 936.587062] env[65121]: DEBUG nova.network.neutron [-] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 936.587062] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 936.587062] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 936.587251] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 936.633687] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106834, 'name': CreateSnapshot_Task, 'duration_secs': 0.869996} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.633977] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 936.634904] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be568b86-17c3-43f2-ba6f-b8afb00d1e95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.661639] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 936.678242] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.703573] env[65121]: DEBUG nova.network.neutron [-] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 936.791931] env[65121]: DEBUG nova.network.neutron [-] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 936.800928] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.734s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.801162] env[65121]: DEBUG nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 936.803884] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.397s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.805564] env[65121]: INFO nova.compute.claims [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.963921] env[65121]: DEBUG nova.network.neutron [-] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 936.979107] env[65121]: DEBUG oslo_vmware.api [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106835, 'name': PowerOnVM_Task, 'duration_secs': 0.608901} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.979486] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.979644] env[65121]: INFO nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Took 8.52 seconds to spawn the instance on the hypervisor. [ 936.979818] env[65121]: DEBUG nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 936.980851] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb464697-b640-430e-8a91-6d76b8b252cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.028207] env[65121]: DEBUG nova.compute.manager [req-c7c64282-6744-4f5f-bc52-c8fb7e3494e6 req-69d0e9d7-700e-4ebb-94ff-4eaca0076a43 service nova] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Received event network-vif-deleted-265542f8-48b6-48e7-b0c5-2d371a74f6da {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 937.028472] env[65121]: DEBUG nova.compute.manager [req-c7c64282-6744-4f5f-bc52-c8fb7e3494e6 req-69d0e9d7-700e-4ebb-94ff-4eaca0076a43 service nova] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Received event network-vif-deleted-b2c8d776-fc6a-4893-a904-f33386dc0317 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 937.068259] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6b30316-5405-4f9c-bcc8-0913d9ea49ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.079589] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d3f118-5598-43b6-9190-b1d0c44d3323 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.121409] env[65121]: DEBUG nova.compute.manager [req-34fb7fab-9467-4d70-bf94-b2a30ccea515 req-73791ea8-4cfb-4af2-84e1-7abd71555856 service nova] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Detach interface failed, port_id=1740aa3b-2428-4bc3-8888-c4edc0734c60, reason: Instance ab41ae41-e69d-47fb-a31a-16748fd3b0f2 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 937.155092] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 937.155451] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3df56cb0-4841-4ee4-9452-d980a305fe5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.167040] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 937.167040] env[65121]: value = "task-5106838" [ 937.167040] env[65121]: _type = "Task" [ 937.167040] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.177742] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106838, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.206569] env[65121]: INFO nova.compute.manager [-] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Took 1.33 seconds to deallocate network for instance. [ 937.299049] env[65121]: INFO nova.compute.manager [-] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Took 1.39 seconds to deallocate network for instance. [ 937.311128] env[65121]: DEBUG nova.compute.utils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 937.317766] env[65121]: DEBUG nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 937.318858] env[65121]: DEBUG nova.network.neutron [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 937.318858] env[65121]: WARNING neutronclient.v2_0.client [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.318858] env[65121]: WARNING neutronclient.v2_0.client [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.320189] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.320189] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 937.408138] env[65121]: DEBUG nova.policy [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd148efb6c7814df182782805ba0e326b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48809b7bf36c4315b9ac315d6b9ca0c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 937.431867] env[65121]: DEBUG nova.network.neutron [-] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 937.466437] env[65121]: INFO nova.compute.manager [-] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Took 1.55 seconds to deallocate network for instance. [ 937.505345] env[65121]: INFO nova.compute.manager [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Took 46.16 seconds to build instance. [ 937.682219] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106838, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.714013] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.751037] env[65121]: DEBUG nova.network.neutron [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Successfully created port: 04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 937.805298] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.817275] env[65121]: DEBUG nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 937.864987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "600d5554-f52a-48ca-941d-1a755d086823" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.864987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.864987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "600d5554-f52a-48ca-941d-1a755d086823-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.864987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.864987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.867979] env[65121]: INFO nova.compute.manager [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Terminating instance [ 937.934895] env[65121]: INFO nova.compute.manager [-] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Took 1.35 seconds to deallocate network for instance. [ 937.974389] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.008235] env[65121]: DEBUG oslo_concurrency.lockutils [None req-17db81b4-c1f4-4c9e-8fac-1b7de626218d tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.666s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.185030] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106838, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.329033] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4bf844-bc33-4bd6-a99e-95107b871be0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.337644] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3981c8f-9df9-41c2-9d9d-e7765db30ef6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.372082] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4134afb7-8f64-4c0b-89f3-519274698101 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.376441] env[65121]: DEBUG nova.compute.manager [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 938.376774] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.377725] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a914346-5737-448d-ae99-00d3b0c1b6b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.389736] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ac14f9-382f-4c40-8884-ecf089b0aa30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.393963] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.394599] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-386e2f72-ea97-4f85-aec5-c337a1b70531 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.407115] env[65121]: DEBUG nova.compute.provider_tree [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.442386] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.468450] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.468673] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.468847] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleting the datastore file [datastore1] 600d5554-f52a-48ca-941d-1a755d086823 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.469377] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44e6579b-c846-4d8e-a5cb-c9bff86de242 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.476619] env[65121]: DEBUG oslo_vmware.api [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 938.476619] env[65121]: value = "task-5106840" [ 938.476619] env[65121]: _type = "Task" [ 938.476619] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.486221] env[65121]: DEBUG oslo_vmware.api [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106840, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.626817] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "ac001ee6-f601-48f7-af0d-42c2387f3c98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.627209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.627446] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "ac001ee6-f601-48f7-af0d-42c2387f3c98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.627641] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.627813] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.630257] env[65121]: INFO nova.compute.manager [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Terminating instance [ 938.682213] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106838, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.837104] env[65121]: DEBUG nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 938.855638] env[65121]: DEBUG nova.compute.manager [req-d779a7cc-7c69-4e43-9ac1-5039caebb6dc req-221e24bc-7794-41a6-822a-fb0400773c17 service nova] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Received event network-vif-deleted-00cda2d7-d8b7-4d99-9030-9f175852e4d9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 938.871830] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 938.871830] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 938.871830] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 938.872124] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 938.872124] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 938.872195] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 938.872353] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.872505] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 938.872663] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 938.872852] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 938.873041] env[65121]: DEBUG nova.virt.hardware [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 938.874288] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07133219-cad3-44c8-917b-fb2209cd3d9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.885657] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0eb7da-906e-46be-b72a-80e09807aa5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.910625] env[65121]: DEBUG nova.scheduler.client.report [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 938.988829] env[65121]: DEBUG oslo_vmware.api [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5106840, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.270086} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.989146] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.989362] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.989562] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.989759] env[65121]: INFO nova.compute.manager [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Took 0.61 seconds to destroy the instance on the hypervisor. [ 938.990257] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 938.990374] env[65121]: DEBUG nova.compute.manager [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 938.990413] env[65121]: DEBUG nova.network.neutron [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 938.990731] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 938.991311] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 938.991609] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.034513] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.134328] env[65121]: DEBUG nova.compute.manager [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 939.134582] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.135931] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f873fdfe-5223-406a-aadb-d7e133ce769d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.145387] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.146467] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91d2291d-5f36-48e9-ab1f-7e255f5038cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.155048] env[65121]: DEBUG oslo_vmware.api [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 939.155048] env[65121]: value = "task-5106841" [ 939.155048] env[65121]: _type = "Task" [ 939.155048] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.166767] env[65121]: DEBUG oslo_vmware.api [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106841, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.181602] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106838, 'name': CloneVM_Task, 'duration_secs': 1.987668} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.181892] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Created linked-clone VM from snapshot [ 939.182894] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8abbd1-318c-4d84-be05-2877a3b11437 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.192462] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Uploading image 91345b02-0e33-4bb1-bc26-4933119dd392 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 939.221208] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 939.221208] env[65121]: value = "vm-993482" [ 939.221208] env[65121]: _type = "VirtualMachine" [ 939.221208] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 939.221894] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3e8badc2-50a3-4f3e-803b-7cf4a0ac2163 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.231836] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease: (returnval){ [ 939.231836] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5205d4ae-a53a-da6b-cc3e-7baef928d5f0" [ 939.231836] env[65121]: _type = "HttpNfcLease" [ 939.231836] env[65121]: } obtained for exporting VM: (result){ [ 939.231836] env[65121]: value = "vm-993482" [ 939.231836] env[65121]: _type = "VirtualMachine" [ 939.231836] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 939.232212] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the lease: (returnval){ [ 939.232212] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5205d4ae-a53a-da6b-cc3e-7baef928d5f0" [ 939.232212] env[65121]: _type = "HttpNfcLease" [ 939.232212] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 939.240765] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.240765] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5205d4ae-a53a-da6b-cc3e-7baef928d5f0" [ 939.240765] env[65121]: _type = "HttpNfcLease" [ 939.240765] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 939.379291] env[65121]: DEBUG nova.network.neutron [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Successfully updated port: 04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 939.415494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.416195] env[65121]: DEBUG nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 939.419388] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.607s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.419614] env[65121]: DEBUG nova.objects.instance [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lazy-loading 'resources' on Instance uuid b22ccce7-c54a-4577-9de0-1fd9c10cd189 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.633014] env[65121]: INFO nova.compute.manager [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Rebuilding instance [ 939.672547] env[65121]: DEBUG nova.compute.manager [req-bf0d3b9c-b150-4deb-8c3e-f083f8d7bb46 req-4c10244f-8436-4888-8ba3-064d0ad568da service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Received event network-vif-deleted-c2c37a16-3107-4510-9c9d-17bf130945ff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 939.672771] env[65121]: INFO nova.compute.manager [req-bf0d3b9c-b150-4deb-8c3e-f083f8d7bb46 req-4c10244f-8436-4888-8ba3-064d0ad568da service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Neutron deleted interface c2c37a16-3107-4510-9c9d-17bf130945ff; detaching it from the instance and deleting it from the info cache [ 939.672935] env[65121]: DEBUG nova.network.neutron [req-bf0d3b9c-b150-4deb-8c3e-f083f8d7bb46 req-4c10244f-8436-4888-8ba3-064d0ad568da service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 939.680064] env[65121]: DEBUG oslo_vmware.api [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106841, 'name': PowerOffVM_Task, 'duration_secs': 0.202764} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.680806] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.680806] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.680806] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-026fb269-c976-46f1-ad9f-0c5cb899ebe1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.692366] env[65121]: DEBUG nova.compute.manager [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 939.693240] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a9ac1e-8c3d-49b0-be12-0cf656afbf80 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.742621] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.742621] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5205d4ae-a53a-da6b-cc3e-7baef928d5f0" [ 939.742621] env[65121]: _type = "HttpNfcLease" [ 939.742621] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 939.742752] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 939.742752] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5205d4ae-a53a-da6b-cc3e-7baef928d5f0" [ 939.742752] env[65121]: _type = "HttpNfcLease" [ 939.742752] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 939.743535] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c5ce85-f267-4322-8fae-d313c7f2c0f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.753389] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5234a0a4-d467-69dd-6f7b-57c0ccccecd1/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 939.753598] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5234a0a4-d467-69dd-6f7b-57c0ccccecd1/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 939.756249] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.756542] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.756754] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Deleting the datastore file [datastore1] ac001ee6-f601-48f7-af0d-42c2387f3c98 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.757932] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8cd323a-50b1-40bd-bfab-732d4dc94c4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.818166] env[65121]: DEBUG oslo_vmware.api [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for the task: (returnval){ [ 939.818166] env[65121]: value = "task-5106844" [ 939.818166] env[65121]: _type = "Task" [ 939.818166] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.825039] env[65121]: DEBUG nova.network.neutron [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 939.828870] env[65121]: DEBUG oslo_vmware.api [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.855036] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-256fa96c-7103-44f6-982e-0ba5c6ec7f6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.884321] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.884568] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.884769] env[65121]: DEBUG nova.network.neutron [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 939.927185] env[65121]: DEBUG nova.compute.utils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 939.928310] env[65121]: DEBUG nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 939.928513] env[65121]: DEBUG nova.network.neutron [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 939.929580] env[65121]: WARNING neutronclient.v2_0.client [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.929580] env[65121]: WARNING neutronclient.v2_0.client [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.932306] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.932306] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.986014] env[65121]: DEBUG nova.policy [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2aba82c33a8f485da007421067986762', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65e498022aee482da65a8176ab61fc71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 940.083784] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Acquiring lock "1b6ad893-e013-4fd5-a829-535d1ea6e001" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.084046] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.085359] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Acquiring lock "1b6ad893-e013-4fd5-a829-535d1ea6e001-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.085359] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.085359] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.088684] env[65121]: INFO nova.compute.manager [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Terminating instance [ 940.177065] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e84b615-6e06-433d-ab1b-568bd52b77f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.188968] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3335c3c-1e8c-424f-b5f3-08b961b691d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.240180] env[65121]: DEBUG nova.compute.manager [req-bf0d3b9c-b150-4deb-8c3e-f083f8d7bb46 req-4c10244f-8436-4888-8ba3-064d0ad568da service nova] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Detach interface failed, port_id=c2c37a16-3107-4510-9c9d-17bf130945ff, reason: Instance 600d5554-f52a-48ca-941d-1a755d086823 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 940.331360] env[65121]: INFO nova.compute.manager [-] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Took 1.34 seconds to deallocate network for instance. [ 940.332507] env[65121]: DEBUG oslo_vmware.api [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Task: {'id': task-5106844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20895} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.337649] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.338817] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.338817] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.338817] env[65121]: INFO nova.compute.manager [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Took 1.20 seconds to destroy the instance on the hypervisor. [ 940.338817] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 940.342486] env[65121]: DEBUG nova.compute.manager [-] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 940.342486] env[65121]: DEBUG nova.network.neutron [-] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 940.342486] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.342737] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.343020] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.385766] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.388899] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.389391] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.405851] env[65121]: DEBUG nova.network.neutron [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Successfully created port: 86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 940.437486] env[65121]: DEBUG nova.network.neutron [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 940.440477] env[65121]: DEBUG nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 940.521239] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b908e5ee-8833-4877-ab39-8530f75cf0be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.530088] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a17607-9541-4d09-8ba9-262f7403edb3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.564387] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9712449-490f-4f9e-9d16-d6d5efa7862f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.573541] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e7ffc6-5267-4c65-a073-d07e98ebe40c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.591695] env[65121]: DEBUG nova.compute.provider_tree [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.595721] env[65121]: DEBUG nova.compute.manager [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 940.595961] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.596418] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2b8a0fe-7bd4-4204-a22e-3be0b29c6c2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.606062] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 940.606062] env[65121]: value = "task-5106845" [ 940.606062] env[65121]: _type = "Task" [ 940.606062] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.619724] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.654485] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.654840] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.711694] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.712321] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1408caa2-0103-4b09-8312-bc221fd400f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.722386] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 940.722386] env[65121]: value = "task-5106846" [ 940.722386] env[65121]: _type = "Task" [ 940.722386] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.737852] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106846, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.842947] env[65121]: WARNING neutronclient.v2_0.client [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.843986] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.844982] env[65121]: WARNING openstack [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.854514] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.996022] env[65121]: DEBUG nova.network.neutron [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 941.118626] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106845, 'name': PowerOffVM_Task, 'duration_secs': 0.247089} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.119423] env[65121]: ERROR nova.scheduler.client.report [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] [req-27058eeb-5196-4fbc-a831-90ec9623946b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-27058eeb-5196-4fbc-a831-90ec9623946b"}]} [ 941.124477] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.124477] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 941.124477] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993416', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'name': 'volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1b6ad893-e013-4fd5-a829-535d1ea6e001', 'attached_at': '', 'detached_at': '', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'serial': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 941.124477] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21764ab-e809-4edb-9248-e73a36dc6398 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.147274] env[65121]: DEBUG nova.network.neutron [-] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 941.150099] env[65121]: DEBUG nova.scheduler.client.report [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 941.153156] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58588c58-7975-4bc0-abb1-e05d9361825e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.162911] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-500f66de-baa9-4c0f-a623-4b5513339969 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.186077] env[65121]: DEBUG nova.scheduler.client.report [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 941.186325] env[65121]: DEBUG nova.compute.provider_tree [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 941.189456] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b592fac-f1ba-4927-9660-d10406715008 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.214256] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] The volume has not been displaced from its original location: [datastore2] volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da/volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 941.222255] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Reconfiguring VM instance instance-00000041 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 941.223753] env[65121]: DEBUG nova.scheduler.client.report [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 941.227115] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08fb2e6c-74b3-44ad-8501-b6113ff780a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.253160] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106846, 'name': PowerOffVM_Task, 'duration_secs': 0.229507} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.254686] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.254938] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.255259] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 941.255259] env[65121]: value = "task-5106847" [ 941.255259] env[65121]: _type = "Task" [ 941.255259] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.256856] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cafe589-4d6b-4e65-95b4-177337806f97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.268930] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.274403] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4394e9b9-a1d2-4180-9e7d-f9f999484af0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.277382] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106847, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.278977] env[65121]: DEBUG nova.scheduler.client.report [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 941.302593] env[65121]: DEBUG nova.compute.manager [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received event network-vif-plugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 941.302593] env[65121]: DEBUG oslo_concurrency.lockutils [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.303059] env[65121]: DEBUG oslo_concurrency.lockutils [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.303248] env[65121]: DEBUG oslo_concurrency.lockutils [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.303357] env[65121]: DEBUG nova.compute.manager [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] No waiting events found dispatching network-vif-plugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 941.303756] env[65121]: WARNING nova.compute.manager [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received unexpected event network-vif-plugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca for instance with vm_state building and task_state spawning. [ 941.303756] env[65121]: DEBUG nova.compute.manager [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received event network-changed-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 941.303868] env[65121]: DEBUG nova.compute.manager [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Refreshing instance network info cache due to event network-changed-04d47dd2-7f48-48e4-aed4-9d693168b8ca. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 941.304309] env[65121]: DEBUG oslo_concurrency.lockutils [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Acquiring lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.371504] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.371933] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.372242] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleting the datastore file [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.375548] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d36179cd-051f-4703-84f5-7c7163e97d5e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.384246] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 941.384246] env[65121]: value = "task-5106849" [ 941.384246] env[65121]: _type = "Task" [ 941.384246] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.398083] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106849, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.452899] env[65121]: DEBUG nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 941.487897] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 941.488335] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 941.488695] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 941.488998] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 941.489219] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 941.490871] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 941.491557] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.491557] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 941.492145] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 941.493791] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 941.493791] env[65121]: DEBUG nova.virt.hardware [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 941.493791] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101e4299-fbdc-4d53-8833-46181b045757 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.502704] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.503718] env[65121]: DEBUG nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Instance network_info: |[{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 941.504831] env[65121]: DEBUG oslo_concurrency.lockutils [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Acquired lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.505288] env[65121]: DEBUG nova.network.neutron [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Refreshing network info cache for port 04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 941.507220] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:95:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04d47dd2-7f48-48e4-aed4-9d693168b8ca', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.517642] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating folder: Project (48809b7bf36c4315b9ac315d6b9ca0c8). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 941.522547] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebfb7b31-0717-45a6-91bf-dad0ef0bbdda {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.529273] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93b133c-0f66-4d56-b655-99087c9871c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.536851] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Created folder: Project (48809b7bf36c4315b9ac315d6b9ca0c8) in parent group-v993268. [ 941.537087] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating folder: Instances. Parent ref: group-v993483. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 941.548746] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2aeac147-5cf0-409e-aeda-e2b8c4234d33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.561610] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Created folder: Instances in parent group-v993483. [ 941.561983] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 941.565506] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.566389] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-12b9233d-5a5e-40b2-9304-328d470c57c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.589856] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.589856] env[65121]: value = "task-5106852" [ 941.589856] env[65121]: _type = "Task" [ 941.589856] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.599745] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106852, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.654053] env[65121]: INFO nova.compute.manager [-] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Took 1.31 seconds to deallocate network for instance. [ 941.774495] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106847, 'name': ReconfigVM_Task, 'duration_secs': 0.276659} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.774922] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Reconfigured VM instance instance-00000041 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 941.783284] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d053a76-3c36-4043-8943-a4f281d35b89 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.800121] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 941.800121] env[65121]: value = "task-5106853" [ 941.800121] env[65121]: _type = "Task" [ 941.800121] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.812411] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106853, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.858021] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4327657-014d-411e-bcc6-5a5ae01d0746 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.868257] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb72dfe1-783d-4ef5-91b7-4b3dbc919250 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.874298] env[65121]: DEBUG nova.compute.manager [req-51fde927-a731-4f7c-9982-ac3ca6a1a0db req-e12dd14e-9d09-4ae6-895c-6a220f06e65f service nova] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Received event network-vif-deleted-afec16fe-3481-4c2e-b503-28fd3f27b085 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 941.909408] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2522222d-ed99-45ec-8d67-6971cb825c3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.920742] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106849, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325162} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.922242] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb46e62a-6e1b-47ce-81f8-432c9005e72f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.927285] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.927991] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.927991] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.943298] env[65121]: DEBUG nova.compute.provider_tree [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 942.019913] env[65121]: WARNING neutronclient.v2_0.client [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.022046] env[65121]: WARNING openstack [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.022046] env[65121]: WARNING openstack [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.068613] env[65121]: DEBUG nova.network.neutron [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Successfully updated port: 86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 942.104233] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106852, 'name': CreateVM_Task, 'duration_secs': 0.376989} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.104455] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.104984] env[65121]: WARNING neutronclient.v2_0.client [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.105487] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.105664] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.106075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 942.106415] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a27368e2-cdb9-42d4-bbe2-7f1c81f698f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.112945] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 942.112945] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ad80c2-a5a6-f34f-a181-0d0f1c6dad90" [ 942.112945] env[65121]: _type = "Task" [ 942.112945] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.124668] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ad80c2-a5a6-f34f-a181-0d0f1c6dad90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.156194] env[65121]: WARNING openstack [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.156194] env[65121]: WARNING openstack [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.166479] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.245889] env[65121]: WARNING neutronclient.v2_0.client [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.246631] env[65121]: WARNING openstack [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.247007] env[65121]: WARNING openstack [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.312724] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106853, 'name': ReconfigVM_Task, 'duration_secs': 0.197896} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.313253] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993416', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'name': 'volume-acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1b6ad893-e013-4fd5-a829-535d1ea6e001', 'attached_at': '', 'detached_at': '', 'volume_id': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da', 'serial': 'acaa084b-972e-4290-a0b1-dfd2ae6ff4da'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 942.313355] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.314222] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad35281-9906-43d1-8478-895ff40f7943 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.323057] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.323207] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-497af914-579f-43ad-b6ea-0b7875218abb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.338325] env[65121]: DEBUG nova.network.neutron [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updated VIF entry in instance network info cache for port 04d47dd2-7f48-48e4-aed4-9d693168b8ca. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 942.338690] env[65121]: DEBUG nova.network.neutron [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 942.403414] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.403864] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.404174] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Deleting the datastore file [datastore2] 1b6ad893-e013-4fd5-a829-535d1ea6e001 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.404503] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b40805c-50f6-4568-b285-11b0c5cb5f97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.412036] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for the task: (returnval){ [ 942.412036] env[65121]: value = "task-5106855" [ 942.412036] env[65121]: _type = "Task" [ 942.412036] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.422180] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106855, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.481810] env[65121]: DEBUG nova.scheduler.client.report [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 942.482099] env[65121]: DEBUG nova.compute.provider_tree [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 103 to 104 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 942.482278] env[65121]: DEBUG nova.compute.provider_tree [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 942.572410] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.572599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquired lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.572777] env[65121]: DEBUG nova.network.neutron [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 942.625894] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ad80c2-a5a6-f34f-a181-0d0f1c6dad90, 'name': SearchDatastore_Task, 'duration_secs': 0.017277} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.626394] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.626677] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.626953] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.627124] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.627310] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.627617] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1265047b-3c8c-4cb2-b53f-b17bab33124b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.641564] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.641773] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.642542] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-159cf815-bf3f-480b-96f4-51d19f1316e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.648871] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 942.648871] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52793a36-2f18-4927-ac0f-1327cd2aed2f" [ 942.648871] env[65121]: _type = "Task" [ 942.648871] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.658026] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52793a36-2f18-4927-ac0f-1327cd2aed2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.842493] env[65121]: DEBUG oslo_concurrency.lockutils [req-92a6ec6b-e5d4-4c98-a06a-eb377853aa72 req-96ef65d6-e364-4d8b-998f-40730c499195 service nova] Releasing lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.923469] env[65121]: DEBUG oslo_vmware.api [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Task: {'id': task-5106855, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130569} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.923706] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.923879] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 942.924061] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 942.924597] env[65121]: INFO nova.compute.manager [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Took 2.33 seconds to destroy the instance on the hypervisor. [ 942.924597] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 942.924713] env[65121]: DEBUG nova.compute.manager [-] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 942.924778] env[65121]: DEBUG nova.network.neutron [-] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 942.925132] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.925536] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.925834] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.964398] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 942.964643] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 942.964793] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 942.964968] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 942.965120] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 942.965516] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 942.965836] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.966173] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 942.966346] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 942.966508] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 942.966720] env[65121]: DEBUG nova.virt.hardware [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 942.967647] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e29900d-d7ea-4d6b-82c3-8e10c045d4eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.971579] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.980168] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80261b94-c363-4b3d-8615-48031414697b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.987575] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.568s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.999422] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.552s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.999690] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.002593] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.039s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.002813] env[65121]: DEBUG nova.objects.instance [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 943.006936] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:23:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '920563c9-c663-4389-9348-0d9ccbca5fea', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.015202] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 943.017084] env[65121]: INFO nova.scheduler.client.report [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Deleted allocations for instance b22ccce7-c54a-4577-9de0-1fd9c10cd189 [ 943.025162] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.025342] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c8685ad-84e7-46a5-a765-23fae9cacdd1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.047585] env[65121]: INFO nova.compute.manager [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Rebuilding instance [ 943.054265] env[65121]: INFO nova.scheduler.client.report [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Deleted allocations for instance 293f93f2-c01d-42c8-b1a7-3056805c77de [ 943.058989] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.058989] env[65121]: value = "task-5106856" [ 943.058989] env[65121]: _type = "Task" [ 943.058989] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.070708] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106856, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.076344] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.078512] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.104159] env[65121]: DEBUG nova.compute.manager [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 943.105516] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701164ea-6c3e-4a8d-9576-7d6d082cad63 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.157809] env[65121]: DEBUG nova.network.neutron [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 943.166141] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52793a36-2f18-4927-ac0f-1327cd2aed2f, 'name': SearchDatastore_Task, 'duration_secs': 0.019667} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.166323] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88710fcb-01f9-4db9-89ad-eaf7014c76e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.173356] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 943.173356] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5249b321-3004-ae94-3b50-9f87ffff3f64" [ 943.173356] env[65121]: _type = "Task" [ 943.173356] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.184557] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5249b321-3004-ae94-3b50-9f87ffff3f64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.228549] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.228951] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.321642] env[65121]: WARNING neutronclient.v2_0.client [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.322014] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.322408] env[65121]: WARNING openstack [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.343123] env[65121]: DEBUG nova.compute.manager [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Received event network-vif-plugged-86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 943.343541] env[65121]: DEBUG oslo_concurrency.lockutils [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Acquiring lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.343613] env[65121]: DEBUG oslo_concurrency.lockutils [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.344180] env[65121]: DEBUG oslo_concurrency.lockutils [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.344180] env[65121]: DEBUG nova.compute.manager [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] No waiting events found dispatching network-vif-plugged-86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 943.344569] env[65121]: WARNING nova.compute.manager [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Received unexpected event network-vif-plugged-86149f88-993f-45cf-94be-cb2781e74ac9 for instance with vm_state building and task_state spawning. [ 943.344736] env[65121]: DEBUG nova.compute.manager [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Received event network-changed-86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 943.345270] env[65121]: DEBUG nova.compute.manager [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Refreshing instance network info cache due to event network-changed-86149f88-993f-45cf-94be-cb2781e74ac9. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 943.345270] env[65121]: DEBUG oslo_concurrency.lockutils [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Acquiring lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.471497] env[65121]: DEBUG nova.network.neutron [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Updating instance_info_cache with network_info: [{"id": "86149f88-993f-45cf-94be-cb2781e74ac9", "address": "fa:16:3e:e4:05:3a", "network": {"id": "9bd97a6f-a92f-49c4-a30f-c317253f98be", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1625344822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65e498022aee482da65a8176ab61fc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86149f88-99", "ovs_interfaceid": "86149f88-993f-45cf-94be-cb2781e74ac9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 943.529106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f187ee6d-11a8-42b3-a648-3cceba01ccc7 tempest-ServersWithSpecificFlavorTestJSON-1330520152 tempest-ServersWithSpecificFlavorTestJSON-1330520152-project-member] Lock "b22ccce7-c54a-4577-9de0-1fd9c10cd189" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.980s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.566637] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f847cbcc-5ff6-4c18-8323-c5dec3896898 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "293f93f2-c01d-42c8-b1a7-3056805c77de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.175s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.574611] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106856, 'name': CreateVM_Task, 'duration_secs': 0.375286} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.574832] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.575418] env[65121]: WARNING neutronclient.v2_0.client [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 943.575885] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.576414] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.576494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 943.576775] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad5975dd-d3fe-4653-a1bf-f539adaa246b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.582957] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 943.582957] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525aa311-dd23-863b-5c3f-316574dfe193" [ 943.582957] env[65121]: _type = "Task" [ 943.582957] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.593402] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525aa311-dd23-863b-5c3f-316574dfe193, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.685329] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5249b321-3004-ae94-3b50-9f87ffff3f64, 'name': SearchDatastore_Task, 'duration_secs': 0.013525} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.685685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.685968] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.686410] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e11a362-1c9a-440c-b344-0c8eed35d5ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.694381] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 943.694381] env[65121]: value = "task-5106857" [ 943.694381] env[65121]: _type = "Task" [ 943.694381] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.708371] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106857, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.950925] env[65121]: DEBUG nova.compute.manager [req-464c251d-aeaf-4ad0-addc-c386da27da0c req-f4cbe934-6887-403f-ba9b-c8409d224bbc service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Received event network-vif-deleted-88b9fcb9-e39c-4751-8499-0f40ea617a1c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 943.951244] env[65121]: INFO nova.compute.manager [req-464c251d-aeaf-4ad0-addc-c386da27da0c req-f4cbe934-6887-403f-ba9b-c8409d224bbc service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Neutron deleted interface 88b9fcb9-e39c-4751-8499-0f40ea617a1c; detaching it from the instance and deleting it from the info cache [ 943.951423] env[65121]: DEBUG nova.network.neutron [req-464c251d-aeaf-4ad0-addc-c386da27da0c req-f4cbe934-6887-403f-ba9b-c8409d224bbc service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 943.973566] env[65121]: DEBUG nova.network.neutron [-] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 943.975553] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Releasing lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.976819] env[65121]: DEBUG nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Instance network_info: |[{"id": "86149f88-993f-45cf-94be-cb2781e74ac9", "address": "fa:16:3e:e4:05:3a", "network": {"id": "9bd97a6f-a92f-49c4-a30f-c317253f98be", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1625344822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65e498022aee482da65a8176ab61fc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86149f88-99", "ovs_interfaceid": "86149f88-993f-45cf-94be-cb2781e74ac9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 943.977100] env[65121]: DEBUG oslo_concurrency.lockutils [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Acquired lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.977806] env[65121]: DEBUG nova.network.neutron [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Refreshing network info cache for port 86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 943.978785] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:05:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b9aabc7c-0f6c-42eb-bd27-493a1496c0c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86149f88-993f-45cf-94be-cb2781e74ac9', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.988722] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Creating folder: Project (65e498022aee482da65a8176ab61fc71). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 943.990613] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ae49bf2-ef42-4c02-b9ef-220af120086f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.010675] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Created folder: Project (65e498022aee482da65a8176ab61fc71) in parent group-v993268. [ 944.011034] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Creating folder: Instances. Parent ref: group-v993487. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 944.012194] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9ab24979-8380-4c28-aba2-8ff78ebed75d tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.013359] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f839ba49-0fcc-40bc-a87d-d17f855d7d34 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.015868] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.697s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.017536] env[65121]: INFO nova.compute.claims [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.033516] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Created folder: Instances in parent group-v993487. [ 944.033701] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 944.034678] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 944.034799] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8240207-e6dc-4eb7-aa07-640cd00336a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.062076] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.062076] env[65121]: value = "task-5106860" [ 944.062076] env[65121]: _type = "Task" [ 944.062076] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.073618] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106860, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.102034] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525aa311-dd23-863b-5c3f-316574dfe193, 'name': SearchDatastore_Task, 'duration_secs': 0.013758} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.102034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.102034] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.102034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.102034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.102034] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.102034] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4931956-f6a8-4674-b76a-c2e6f0878296 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.124080] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.124468] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cae417a-5041-43f7-98c5-4c1e370c35ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.130875] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.132054] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 944.132054] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b0e4060-f6e1-4f75-a334-d2a6cbcce3c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.136902] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 944.136902] env[65121]: value = "task-5106861" [ 944.136902] env[65121]: _type = "Task" [ 944.136902] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.143684] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 944.143684] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52262278-1981-39e0-0c4f-e2711290fd88" [ 944.143684] env[65121]: _type = "Task" [ 944.143684] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.153242] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.161742] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52262278-1981-39e0-0c4f-e2711290fd88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.214053] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106857, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.455224] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f0f77aa-a131-4648-b871-191160baf6bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.467370] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a433509-4b74-4ba6-8097-fcf80699027d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.479348] env[65121]: INFO nova.compute.manager [-] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Took 1.55 seconds to deallocate network for instance. [ 944.491148] env[65121]: WARNING neutronclient.v2_0.client [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.491923] env[65121]: WARNING openstack [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.492382] env[65121]: WARNING openstack [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.517454] env[65121]: DEBUG nova.compute.manager [req-464c251d-aeaf-4ad0-addc-c386da27da0c req-f4cbe934-6887-403f-ba9b-c8409d224bbc service nova] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Detach interface failed, port_id=88b9fcb9-e39c-4751-8499-0f40ea617a1c, reason: Instance 1b6ad893-e013-4fd5-a829-535d1ea6e001 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 944.574141] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106860, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.650060] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106861, 'name': PowerOffVM_Task, 'duration_secs': 0.373977} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.653913] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.654473] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.657667] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f85d52e-6140-46d5-b786-45aced49edc9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.668243] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52262278-1981-39e0-0c4f-e2711290fd88, 'name': SearchDatastore_Task, 'duration_secs': 0.058713} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.670907] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 944.671054] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc55479f-48ea-4008-adfb-c643677beac1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.673748] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d40e485b-51b5-41b4-b572-2973b61d096f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.679501] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 944.679501] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cf391e-ea30-babe-4f65-e025e1a1a6d2" [ 944.679501] env[65121]: _type = "Task" [ 944.679501] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.681130] env[65121]: WARNING openstack [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.681386] env[65121]: WARNING openstack [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.698471] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cf391e-ea30-babe-4f65-e025e1a1a6d2, 'name': SearchDatastore_Task, 'duration_secs': 0.0146} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.701858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.702169] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.702525] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-192fef3d-bcc4-48a4-b78a-d0ca85b4f441 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.711761] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106857, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.626182} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.713364] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.713650] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.713989] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 944.713989] env[65121]: value = "task-5106863" [ 944.713989] env[65121]: _type = "Task" [ 944.713989] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.714219] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55057708-9915-48a2-82b1-16f8503e8011 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.730207] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.731939] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 944.731939] env[65121]: value = "task-5106864" [ 944.731939] env[65121]: _type = "Task" [ 944.731939] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.748025] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 944.748546] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 944.748931] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleting the datastore file [datastore2] 274d0ccd-c707-4a68-b280-16de2bc74d73 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.749348] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c167745-2882-48fa-84f2-5782de11dd6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.758130] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106864, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.765494] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 944.765494] env[65121]: value = "task-5106865" [ 944.765494] env[65121]: _type = "Task" [ 944.765494] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.776433] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106865, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.790961] env[65121]: WARNING neutronclient.v2_0.client [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.791815] env[65121]: WARNING openstack [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.792284] env[65121]: WARNING openstack [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.896174] env[65121]: DEBUG nova.network.neutron [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Updated VIF entry in instance network info cache for port 86149f88-993f-45cf-94be-cb2781e74ac9. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 944.896712] env[65121]: DEBUG nova.network.neutron [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Updating instance_info_cache with network_info: [{"id": "86149f88-993f-45cf-94be-cb2781e74ac9", "address": "fa:16:3e:e4:05:3a", "network": {"id": "9bd97a6f-a92f-49c4-a30f-c317253f98be", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1625344822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65e498022aee482da65a8176ab61fc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86149f88-99", "ovs_interfaceid": "86149f88-993f-45cf-94be-cb2781e74ac9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 945.043189] env[65121]: INFO nova.compute.manager [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Took 0.56 seconds to detach 1 volumes for instance. [ 945.047116] env[65121]: DEBUG nova.compute.manager [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Deleting volume: acaa084b-972e-4290-a0b1-dfd2ae6ff4da {{(pid=65121) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 945.080918] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106860, 'name': CreateVM_Task, 'duration_secs': 0.6008} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.085537] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.085537] env[65121]: WARNING neutronclient.v2_0.client [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.085833] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.085972] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.086363] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 945.086710] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff2026d7-a6f1-44e9-99b3-619a2837b4cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.095822] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 945.095822] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522ef650-8ff2-b514-0a5c-2f88db7f2bfc" [ 945.095822] env[65121]: _type = "Task" [ 945.095822] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.118606] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522ef650-8ff2-b514-0a5c-2f88db7f2bfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.231418] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106863, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.248479] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106864, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075437} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.252758] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.254060] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6644dd-c6f7-4335-a66d-7515881e06db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.281924] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.288342] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd1a7110-90d8-4cf3-b885-57419301d809 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.317396] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106865, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.313852} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.319396] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.319662] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.319903] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.322874] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 945.322874] env[65121]: value = "task-5106867" [ 945.322874] env[65121]: _type = "Task" [ 945.322874] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.338238] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106867, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.399256] env[65121]: DEBUG oslo_concurrency.lockutils [req-5788fdfc-6439-47ab-90d9-411ed1780510 req-cb2ad4df-fe85-43e6-a0bb-db61c69df8c0 service nova] Releasing lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.440480] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.440751] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 945.440960] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.441165] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 945.441335] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 945.444435] env[65121]: INFO nova.compute.manager [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Terminating instance [ 945.587957] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48858f3f-b609-47c9-a067-2978e21e6c94 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.601103] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40be392f-c28c-48bb-bba3-c047180614af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.613191] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.613557] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522ef650-8ff2-b514-0a5c-2f88db7f2bfc, 'name': SearchDatastore_Task, 'duration_secs': 0.063068} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.646340] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.647238] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.647654] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.647654] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.647835] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.648430] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e111e509-2975-4b00-b011-89d6804eab1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.653906] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98d7722-62ac-4fff-af71-059f0181fdad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.662783] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7906c3-d4a2-4ccc-8b51-bffc8dfd42e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.667891] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.668201] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.669305] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ddb5e63-0a5c-486b-a641-126d6235ad7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.680148] env[65121]: DEBUG nova.compute.provider_tree [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.685614] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 945.685614] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528b25b4-92ee-8c3f-94d4-fa9f0640aa4b" [ 945.685614] env[65121]: _type = "Task" [ 945.685614] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.695221] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528b25b4-92ee-8c3f-94d4-fa9f0640aa4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.728398] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106863, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641206} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.729339] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.729595] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.730036] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-616ac74e-ce71-4d2f-93ce-a614047218db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.739148] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 945.739148] env[65121]: value = "task-5106868" [ 945.739148] env[65121]: _type = "Task" [ 945.739148] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.749346] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106868, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.840553] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106867, 'name': ReconfigVM_Task, 'duration_secs': 0.324653} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.840784] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.841460] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87d3ffd0-1cf1-4085-9ed4-a992751e6ccc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.849355] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 945.849355] env[65121]: value = "task-5106869" [ 945.849355] env[65121]: _type = "Task" [ 945.849355] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.861092] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106869, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.954423] env[65121]: DEBUG nova.compute.manager [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 945.954661] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.955843] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cf20e9-ccfc-4c64-aa04-179b428b95b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.965220] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.966135] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8408c73b-7ebb-4832-b22e-d01eba9f6a06 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.973987] env[65121]: DEBUG oslo_vmware.api [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 945.973987] env[65121]: value = "task-5106870" [ 945.973987] env[65121]: _type = "Task" [ 945.973987] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.985506] env[65121]: DEBUG oslo_vmware.api [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.184972] env[65121]: DEBUG nova.scheduler.client.report [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.200622] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528b25b4-92ee-8c3f-94d4-fa9f0640aa4b, 'name': SearchDatastore_Task, 'duration_secs': 0.048774} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.201549] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74d82bf4-e06d-4bcd-818a-ecc47e5b150f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.210489] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 946.210489] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52992a1f-7435-1faf-74c6-aed00833cac8" [ 946.210489] env[65121]: _type = "Task" [ 946.210489] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.225474] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52992a1f-7435-1faf-74c6-aed00833cac8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.249998] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106868, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07146} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.250114] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.250997] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ff2b21-54fe-48cb-a633-c656e6f922cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.285407] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.285901] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-592892d3-2f73-4615-9c7d-60e1cbcf05a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.317530] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 946.317530] env[65121]: value = "task-5106871" [ 946.317530] env[65121]: _type = "Task" [ 946.317530] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.330373] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.359398] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106869, 'name': Rename_Task, 'duration_secs': 0.183111} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.361555] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.362496] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff2e58a6-51b9-48d8-8c46-426cefe11267 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.369028] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 946.369330] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 946.369485] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 946.369659] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 946.369794] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 946.369931] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 946.370151] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.370303] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 946.370474] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 946.370610] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 946.370779] env[65121]: DEBUG nova.virt.hardware [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 946.371707] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce5a0e6-7634-46d5-80fa-8286ca582a61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.375946] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 946.375946] env[65121]: value = "task-5106872" [ 946.375946] env[65121]: _type = "Task" [ 946.375946] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.383621] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8611801d-7e84-4be4-89ba-ba19fd57c8bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.392967] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106872, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.406594] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:04:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6966f473-59ac-49bb-9b7a-22c61f4e61e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '748bfc22-93f1-459b-9b59-18583587dd17', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.418358] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 946.418762] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.419124] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be02b08c-7112-49f3-bd98-f1152d556eab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.440997] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.440997] env[65121]: value = "task-5106873" [ 946.440997] env[65121]: _type = "Task" [ 946.440997] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.452569] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106873, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.486680] env[65121]: DEBUG oslo_vmware.api [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106870, 'name': PowerOffVM_Task, 'duration_secs': 0.266605} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.486875] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.487046] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.487306] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9488771-616c-405e-b0bf-a269c1c5acee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.563389] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.564445] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.564445] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Deleting the datastore file [datastore1] 5f6e4f46-0745-42c0-a779-6cffb60e21ca {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.564445] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-faf49fcb-5287-477c-8d36-89af87a1247f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.572593] env[65121]: DEBUG oslo_vmware.api [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for the task: (returnval){ [ 946.572593] env[65121]: value = "task-5106875" [ 946.572593] env[65121]: _type = "Task" [ 946.572593] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.583577] env[65121]: DEBUG oslo_vmware.api [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.695935] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.680s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.696708] env[65121]: DEBUG nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 946.699907] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.410s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.701449] env[65121]: INFO nova.compute.claims [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.725262] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52992a1f-7435-1faf-74c6-aed00833cac8, 'name': SearchDatastore_Task, 'duration_secs': 0.013467} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.725454] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.725683] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 90cb98b1-9520-4caf-b6c4-80abaa7a9221/90cb98b1-9520-4caf-b6c4-80abaa7a9221.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.726160] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-216041fb-a30d-4e07-80bb-94ede5c62e1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.737517] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 946.737517] env[65121]: value = "task-5106876" [ 946.737517] env[65121]: _type = "Task" [ 946.737517] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.655655] env[65121]: DEBUG nova.compute.utils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 947.659680] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.659902] env[65121]: WARNING oslo_vmware.common.loopingcall [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] task run outlasted interval by 0.422111 sec [ 947.674141] env[65121]: DEBUG nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 947.674443] env[65121]: DEBUG nova.network.neutron [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 947.674701] env[65121]: WARNING neutronclient.v2_0.client [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.675058] env[65121]: WARNING neutronclient.v2_0.client [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.675769] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.676171] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.717034] env[65121]: DEBUG oslo_vmware.api [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Task: {'id': task-5106875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.233304} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.717406] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106871, 'name': ReconfigVM_Task, 'duration_secs': 0.389873} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.717686] env[65121]: DEBUG oslo_vmware.api [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106872, 'name': PowerOnVM_Task, 'duration_secs': 0.55622} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.717904] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106873, 'name': CreateVM_Task, 'duration_secs': 0.44801} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.725216] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.725479] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 947.725604] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 947.725704] env[65121]: INFO nova.compute.manager [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Took 1.77 seconds to destroy the instance on the hypervisor. [ 947.726300] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 947.726300] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4/59d6005d-6caf-4898-b791-70d9c015cdb4.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.726935] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.727337] env[65121]: INFO nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Took 8.89 seconds to spawn the instance on the hypervisor. [ 947.727337] env[65121]: DEBUG nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 947.727788] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.727788] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66252} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.728893] env[65121]: DEBUG nova.compute.manager [-] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 947.728893] env[65121]: DEBUG nova.network.neutron [-] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 947.728996] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.729663] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.729714] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.736283] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4935d71f-7b84-4656-a32e-36cf698d03d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.738667] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae77dc42-9611-436b-9919-94ab04908823 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.742240] env[65121]: WARNING neutronclient.v2_0.client [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 947.742620] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.742760] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.743080] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 947.743361] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 90cb98b1-9520-4caf-b6c4-80abaa7a9221/90cb98b1-9520-4caf-b6c4-80abaa7a9221.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.743608] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.747017] env[65121]: DEBUG nova.policy [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7f1d5cb6ec4c0f8a13f836eb712340', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ce7f6698e214d73ae43427601058af8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 947.749060] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb48f048-3d1c-4627-9a08-f3a283634280 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.750757] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dac92fbe-61cd-4c81-9608-949d8b127f9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.762115] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 947.762115] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52218410-01d6-66be-240a-c9a0c4fef26a" [ 947.762115] env[65121]: _type = "Task" [ 947.762115] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.770977] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 947.770977] env[65121]: value = "task-5106878" [ 947.770977] env[65121]: _type = "Task" [ 947.770977] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.771388] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 947.771388] env[65121]: value = "task-5106877" [ 947.771388] env[65121]: _type = "Task" [ 947.771388] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.786176] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52218410-01d6-66be-240a-c9a0c4fef26a, 'name': SearchDatastore_Task, 'duration_secs': 0.018669} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.787396] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.787644] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.787891] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.788134] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.788332] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.791485] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37150a3e-36c4-4b15-af78-a73896a9ec13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.804731] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106877, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.805830] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106878, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.819334] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.819334] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.820413] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14fb57a-df9f-4eba-ace8-70ae11c61633 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.831060] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 947.831060] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523f9158-bda5-6c55-3a82-a1f20e70494a" [ 947.831060] env[65121]: _type = "Task" [ 947.831060] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.838595] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.851432] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523f9158-bda5-6c55-3a82-a1f20e70494a, 'name': SearchDatastore_Task, 'duration_secs': 0.016744} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.852687] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47bc5dad-2987-41ec-91b5-a39dfe1b6e64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.861992] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 947.861992] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529abd09-268e-cf35-3b84-05fd22bb84bc" [ 947.861992] env[65121]: _type = "Task" [ 947.861992] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.874821] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529abd09-268e-cf35-3b84-05fd22bb84bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.161644] env[65121]: DEBUG nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 948.269207] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e662a085-3d96-4f80-96ee-7df4ea58482a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.291337] env[65121]: INFO nova.compute.manager [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Took 40.45 seconds to build instance. [ 948.294619] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75aac02a-a933-4745-b9dd-0019309ff1dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.304670] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106877, 'name': Rename_Task, 'duration_secs': 0.243771} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.305123] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106878, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080721} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.305984] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.305984] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.307145] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02572e30-a184-452d-9ac6-765708fe72a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.309650] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856507c6-c05b-4b85-8c06-bd9c6263ffaa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.342039] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d938c4f-1bc3-441c-a812-bc6bf0d81dfb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.370951] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 90cb98b1-9520-4caf-b6c4-80abaa7a9221/90cb98b1-9520-4caf-b6c4-80abaa7a9221.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.373897] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c842a29-20f6-4559-9321-b50dacae870e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.392725] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 948.392725] env[65121]: value = "task-5106879" [ 948.392725] env[65121]: _type = "Task" [ 948.392725] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.406017] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70b49c2-4754-49e6-a959-5a03b7cb72c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.412350] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 948.412350] env[65121]: value = "task-5106880" [ 948.412350] env[65121]: _type = "Task" [ 948.412350] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.432686] env[65121]: DEBUG nova.compute.provider_tree [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.437173] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106879, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.437173] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529abd09-268e-cf35-3b84-05fd22bb84bc, 'name': SearchDatastore_Task, 'duration_secs': 0.014984} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.437173] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.437173] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.437173] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc49a6ca-c864-4019-b3d0-d354453c62cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.442049] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.449334] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 948.449334] env[65121]: value = "task-5106881" [ 948.449334] env[65121]: _type = "Task" [ 948.449334] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.463914] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106881, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.485793] env[65121]: DEBUG nova.network.neutron [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Successfully created port: e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 948.609452] env[65121]: DEBUG nova.network.neutron [-] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 948.803377] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5bdc68c4-912c-42f1-a138-f4e12bd1480b tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.966s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.913481] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106879, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.925362] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.937078] env[65121]: DEBUG nova.scheduler.client.report [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.952355] env[65121]: DEBUG nova.compute.manager [req-b6586763-2523-4ab8-9041-254042800784 req-14731f5c-ada6-4a01-b889-43bb3c3f1f40 service nova] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Received event network-vif-deleted-fc77c1b5-9d9b-479a-a553-e7961313af08 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 948.965791] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106881, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.116150] env[65121]: INFO nova.compute.manager [-] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Took 1.39 seconds to deallocate network for instance. [ 949.165582] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5234a0a4-d467-69dd-6f7b-57c0ccccecd1/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 949.166757] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c00a2bc-c857-43db-ad85-b9475163de69 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.175419] env[65121]: DEBUG nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 949.177798] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5234a0a4-d467-69dd-6f7b-57c0ccccecd1/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 949.177798] env[65121]: ERROR oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5234a0a4-d467-69dd-6f7b-57c0ccccecd1/disk-0.vmdk due to incomplete transfer. [ 949.180145] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e986488f-108f-487d-92ba-a70a912e25ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.189596] env[65121]: DEBUG oslo_vmware.rw_handles [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5234a0a4-d467-69dd-6f7b-57c0ccccecd1/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 949.189596] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Uploaded image 91345b02-0e33-4bb1-bc26-4933119dd392 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 949.192861] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 949.192861] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3b51f54f-0ad8-4ece-b30f-5fef13fa96f2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.206818] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 949.206818] env[65121]: value = "task-5106882" [ 949.206818] env[65121]: _type = "Task" [ 949.206818] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.209328] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 949.209561] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 949.209702] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 949.210175] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 949.210175] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 949.210175] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 949.210349] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.210553] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 949.210657] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 949.210805] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 949.210964] env[65121]: DEBUG nova.virt.hardware [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 949.212517] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3f3562-9923-483c-8e1f-00a6d5fc141c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.228046] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f06080-c3b7-4906-8bc4-dcdaed801516 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.232538] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106882, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.409558] env[65121]: DEBUG oslo_vmware.api [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106879, 'name': PowerOnVM_Task, 'duration_secs': 0.694103} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.409879] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.410152] env[65121]: DEBUG nova.compute.manager [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 949.411031] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6eddf41-733b-4d9c-8db9-4da645c6a65a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.429431] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106880, 'name': ReconfigVM_Task, 'duration_secs': 0.784896} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.429897] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 90cb98b1-9520-4caf-b6c4-80abaa7a9221/90cb98b1-9520-4caf-b6c4-80abaa7a9221.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.430649] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f83ff61-0d42-44c6-9af5-af938bf51948 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.438785] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 949.438785] env[65121]: value = "task-5106883" [ 949.438785] env[65121]: _type = "Task" [ 949.438785] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.450226] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.750s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.450931] env[65121]: DEBUG nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 949.454621] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106883, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.454621] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.026s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.455471] env[65121]: DEBUG nova.objects.instance [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lazy-loading 'resources' on Instance uuid 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.468253] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106881, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644577} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.469271] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 949.469523] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.469761] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84bb65b8-b950-42aa-a0e6-69ca86b61a62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.478807] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 949.478807] env[65121]: value = "task-5106884" [ 949.478807] env[65121]: _type = "Task" [ 949.478807] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.490653] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106884, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.624421] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.724338] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106882, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.934240] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.960679] env[65121]: DEBUG nova.compute.utils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 949.967113] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106883, 'name': Rename_Task, 'duration_secs': 0.316029} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.968239] env[65121]: DEBUG nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 949.968392] env[65121]: DEBUG nova.network.neutron [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 949.968716] env[65121]: WARNING neutronclient.v2_0.client [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.969027] env[65121]: WARNING neutronclient.v2_0.client [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.969741] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.970241] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.981565] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.982873] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea14a0db-4acd-46e4-bc12-d97aab968c97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.993716] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 949.993716] env[65121]: value = "task-5106885" [ 949.993716] env[65121]: _type = "Task" [ 949.993716] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.997297] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106884, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.246088} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.000492] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.002194] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd68784-c344-4c3e-bdcb-1ab4a999ada8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.016679] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106885, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.039665] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.042669] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7b3a66e-9091-47e6-a02d-77ea97666633 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.065315] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 950.065315] env[65121]: value = "task-5106886" [ 950.065315] env[65121]: _type = "Task" [ 950.065315] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.078066] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106886, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.152432] env[65121]: DEBUG nova.network.neutron [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Successfully updated port: e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 950.221254] env[65121]: DEBUG nova.policy [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4853c778ab79419da0e9de23ee3617a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f056059180af48bba4587006efb0b211', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 950.230488] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106882, 'name': Destroy_Task, 'duration_secs': 0.711291} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.231044] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Destroyed the VM [ 950.231295] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 950.231591] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9afeceb9-7bd9-4024-a738-ab9152ca9f98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.241345] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 950.241345] env[65121]: value = "task-5106887" [ 950.241345] env[65121]: _type = "Task" [ 950.241345] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.255614] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106887, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.469690] env[65121]: DEBUG nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 950.509175] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106885, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.521395] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02485fbc-e42a-4a50-8577-7ee65a42ccbf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.534631] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73506068-f3c9-4846-92e7-3dc7c0da2b17 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.579013] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a941cd2-3989-4261-9a27-cca031cd7f56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.586269] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 950.586497] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.594274] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106886, 'name': ReconfigVM_Task, 'duration_secs': 0.415093} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.596497] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 274d0ccd-c707-4a68-b280-16de2bc74d73/274d0ccd-c707-4a68-b280-16de2bc74d73.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.597219] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c00aac66-e00c-4b46-be5d-c84b763faca5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.600061] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e92ac89-79b6-4d55-bd0a-faefdcc9f00e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.617346] env[65121]: DEBUG nova.compute.provider_tree [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.621577] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 950.621577] env[65121]: value = "task-5106888" [ 950.621577] env[65121]: _type = "Task" [ 950.621577] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.630605] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106888, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.639084] env[65121]: DEBUG nova.network.neutron [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Successfully created port: ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 950.654518] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.654518] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.654672] env[65121]: DEBUG nova.network.neutron [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 950.752593] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106887, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.009890] env[65121]: DEBUG oslo_vmware.api [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106885, 'name': PowerOnVM_Task, 'duration_secs': 0.545177} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.010364] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.010696] env[65121]: INFO nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Took 9.56 seconds to spawn the instance on the hypervisor. [ 951.010992] env[65121]: DEBUG nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 951.012279] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde40d78-5c77-4538-b698-098369f2f4c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.090069] env[65121]: DEBUG nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 951.122383] env[65121]: DEBUG nova.scheduler.client.report [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 951.138713] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106888, 'name': Rename_Task, 'duration_secs': 0.252489} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.138912] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.139259] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e59fc1d-2c69-4ae3-9a8d-2df95bb42fbf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.147834] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 951.147834] env[65121]: value = "task-5106889" [ 951.147834] env[65121]: _type = "Task" [ 951.147834] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.156781] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106889, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.158727] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.159100] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.253440] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106887, 'name': RemoveSnapshot_Task, 'duration_secs': 0.638907} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.253599] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 951.253939] env[65121]: DEBUG nova.compute.manager [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 951.254787] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3cdfe9-55c4-4c7a-991b-a6f2647ea19b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.376664] env[65121]: DEBUG nova.network.neutron [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 951.487715] env[65121]: DEBUG nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 951.524428] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 951.524667] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 951.524814] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 951.524982] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 951.525207] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 951.525335] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 951.525528] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.525750] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 951.525942] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 951.526147] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 951.526315] env[65121]: DEBUG nova.virt.hardware [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 951.527345] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120e738a-4d8c-4b52-b3ca-fed3072dbc22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.540924] env[65121]: INFO nova.compute.manager [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Took 38.16 seconds to build instance. [ 951.553314] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3449807-f364-479f-bea2-04dba14c06b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.631532] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.177s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.634103] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.400s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.648866] env[65121]: INFO nova.compute.claims [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.655416] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.661205] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106889, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.667239] env[65121]: INFO nova.scheduler.client.report [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleted allocations for instance 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec [ 951.768037] env[65121]: INFO nova.compute.manager [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Shelve offloading [ 952.053224] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c17babe-e875-4030-8b4c-f5b937f08789 tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.678s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.160303] env[65121]: DEBUG oslo_vmware.api [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106889, 'name': PowerOnVM_Task, 'duration_secs': 0.889996} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.161319] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.161519] env[65121]: DEBUG nova.compute.manager [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 952.162746] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae93cdcb-8d7d-475a-80eb-6bf9d907c7b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.180255] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a44580dc-0ce4-425f-a133-4c8a38ab7699 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "30b01b2b-2cbf-4c34-86cd-529a95a9c3ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.181s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.272513] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.273021] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61ee360d-f999-4e13-be38-b8faccedee05 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.282837] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 952.282837] env[65121]: value = "task-5106890" [ 952.282837] env[65121]: _type = "Task" [ 952.282837] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.287181] env[65121]: DEBUG nova.network.neutron [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Successfully updated port: ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 952.294303] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 952.295430] env[65121]: DEBUG nova.compute.manager [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 952.297156] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba1b1b8-cd15-4fb4-b8c1-60af9f77caa5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.303627] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.304192] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.313363] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.313529] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.313694] env[65121]: DEBUG nova.network.neutron [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 952.431388] env[65121]: WARNING neutronclient.v2_0.client [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.435023] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.435023] env[65121]: WARNING openstack [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.472105] env[65121]: DEBUG nova.compute.manager [req-4484b64d-09ba-4040-910d-bdc7ec449e07 req-d25bc50c-bc5e-468a-bfc9-f37b35e1baf3 service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Received event network-vif-plugged-e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 952.472319] env[65121]: DEBUG oslo_concurrency.lockutils [req-4484b64d-09ba-4040-910d-bdc7ec449e07 req-d25bc50c-bc5e-468a-bfc9-f37b35e1baf3 service nova] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.472510] env[65121]: DEBUG oslo_concurrency.lockutils [req-4484b64d-09ba-4040-910d-bdc7ec449e07 req-d25bc50c-bc5e-468a-bfc9-f37b35e1baf3 service nova] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.472669] env[65121]: DEBUG oslo_concurrency.lockutils [req-4484b64d-09ba-4040-910d-bdc7ec449e07 req-d25bc50c-bc5e-468a-bfc9-f37b35e1baf3 service nova] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.472845] env[65121]: DEBUG nova.compute.manager [req-4484b64d-09ba-4040-910d-bdc7ec449e07 req-d25bc50c-bc5e-468a-bfc9-f37b35e1baf3 service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] No waiting events found dispatching network-vif-plugged-e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 952.473131] env[65121]: WARNING nova.compute.manager [req-4484b64d-09ba-4040-910d-bdc7ec449e07 req-d25bc50c-bc5e-468a-bfc9-f37b35e1baf3 service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Received unexpected event network-vif-plugged-e31cbeae-05bc-416d-9df3-10f09d947ba4 for instance with vm_state building and task_state spawning. [ 952.525307] env[65121]: DEBUG nova.network.neutron [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating instance_info_cache with network_info: [{"id": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "address": "fa:16:3e:0e:bf:f3", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape31cbeae-05", "ovs_interfaceid": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.692633] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.793367] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.797242] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.797242] env[65121]: DEBUG nova.network.neutron [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 952.818791] env[65121]: WARNING neutronclient.v2_0.client [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.819628] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.820058] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.968326] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.968326] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.035498] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.035899] env[65121]: DEBUG nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Instance network_info: |[{"id": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "address": "fa:16:3e:0e:bf:f3", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape31cbeae-05", "ovs_interfaceid": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 953.036580] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:bf:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e31cbeae-05bc-416d-9df3-10f09d947ba4', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.046916] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Creating folder: Project (1ce7f6698e214d73ae43427601058af8). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 953.046916] env[65121]: WARNING neutronclient.v2_0.client [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.047132] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.047569] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.056126] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-704e432f-6e01-41cf-a4f5-e7b00e9f5bed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.076608] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Created folder: Project (1ce7f6698e214d73ae43427601058af8) in parent group-v993268. [ 953.076869] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Creating folder: Instances. Parent ref: group-v993491. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 953.077146] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-57deafb2-faa8-4fcd-9a1a-50caec0f9091 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.091631] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Created folder: Instances in parent group-v993491. [ 953.091904] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 953.092377] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 953.092617] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d090a92-890d-42cc-988c-2a3bc0dc8b83 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.130953] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.130953] env[65121]: value = "task-5106893" [ 953.130953] env[65121]: _type = "Task" [ 953.130953] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.141114] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106893, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.166929] env[65121]: DEBUG nova.network.neutron [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e0942f-50", "ovs_interfaceid": "71e0942f-5026-4128-ba81-16311feb9b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 953.170162] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbf7f79-42b2-46be-93ad-7666fd69de8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.181043] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab6e64d-5c6e-417f-af7b-6953089cc8bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.226260] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fbfb12-7abe-412a-bbe7-38bc7d59201e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.236192] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0bb08c-adb7-4373-99db-715a87f5cecb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.254365] env[65121]: DEBUG nova.compute.provider_tree [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.300038] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.300234] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.349026] env[65121]: DEBUG nova.network.neutron [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 953.376584] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.376584] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.460941] env[65121]: WARNING neutronclient.v2_0.client [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.460941] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.460941] env[65121]: WARNING openstack [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.551596] env[65121]: DEBUG nova.network.neutron [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Updating instance_info_cache with network_info: [{"id": "ad549932-0308-4165-96b1-9b4e419b6d05", "address": "fa:16:3e:78:7d:b1", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad549932-03", "ovs_interfaceid": "ad549932-0308-4165-96b1-9b4e419b6d05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 953.643085] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106893, 'name': CreateVM_Task, 'duration_secs': 0.408477} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.643384] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.644017] env[65121]: WARNING neutronclient.v2_0.client [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.645911] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.645911] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.645911] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 953.645911] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffa14f85-20b5-4c6c-9b44-f8ca8e5145ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.650683] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 953.650683] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52eeb482-e5dc-fcea-4142-3d6670bc23f5" [ 953.650683] env[65121]: _type = "Task" [ 953.650683] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.659816] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52eeb482-e5dc-fcea-4142-3d6670bc23f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.673631] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.674098] env[65121]: WARNING neutronclient.v2_0.client [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.674790] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.675455] env[65121]: WARNING openstack [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.680562] env[65121]: WARNING neutronclient.v2_0.client [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.758552] env[65121]: DEBUG nova.scheduler.client.report [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.056876] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.056876] env[65121]: DEBUG nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Instance network_info: |[{"id": "ad549932-0308-4165-96b1-9b4e419b6d05", "address": "fa:16:3e:78:7d:b1", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad549932-03", "ovs_interfaceid": "ad549932-0308-4165-96b1-9b4e419b6d05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 954.056876] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:7d:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad549932-0308-4165-96b1-9b4e419b6d05', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.067894] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Creating folder: Project (f056059180af48bba4587006efb0b211). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 954.068549] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2db978bc-7b77-45d9-b452-732641a69d08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.083552] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Created folder: Project (f056059180af48bba4587006efb0b211) in parent group-v993268. [ 954.083934] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Creating folder: Instances. Parent ref: group-v993494. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 954.084279] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a4b17dc-25d9-4071-988d-7d4789fda228 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.097195] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Created folder: Instances in parent group-v993494. [ 954.097476] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 954.097677] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.097888] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f135d5eb-8c35-4795-a822-2f27146a14b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.116805] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.117931] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d2320d-8674-415e-b1ac-b7b822b02544 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.126926] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.129266] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-603c69e8-9c87-4c3a-98ed-a3caa449f29b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.131025] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.131025] env[65121]: value = "task-5106896" [ 954.131025] env[65121]: _type = "Task" [ 954.131025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.140522] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106896, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.163505] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52eeb482-e5dc-fcea-4142-3d6670bc23f5, 'name': SearchDatastore_Task, 'duration_secs': 0.028207} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.163919] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.164265] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.164608] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.164824] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.165540] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.165540] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3f019c4-38e1-4efd-bc87-78b1bbf52fbd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.178916] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.179219] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.179917] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8f3e44a-4979-4a7f-b5fa-2f3af14940fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.191925] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 954.191925] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5249a149-71db-2882-f56a-6a5c84aa8e8b" [ 954.191925] env[65121]: _type = "Task" [ 954.191925] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.202465] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5249a149-71db-2882-f56a-6a5c84aa8e8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.208516] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.208729] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.208887] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleting the datastore file [datastore2] e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.209351] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d0e8681-69cf-4355-a2ce-363d016268e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.211931] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.212231] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.212505] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.212702] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.212858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.218404] env[65121]: INFO nova.compute.manager [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Terminating instance [ 954.222184] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 954.222184] env[65121]: value = "task-5106898" [ 954.222184] env[65121]: _type = "Task" [ 954.222184] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.238428] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.272142] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.638s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.273151] env[65121]: DEBUG nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 954.276222] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 24.389s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.321449] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "59d6005d-6caf-4898-b791-70d9c015cdb4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.321784] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.322169] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "59d6005d-6caf-4898-b791-70d9c015cdb4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.322416] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.322627] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.326340] env[65121]: INFO nova.compute.manager [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Terminating instance [ 954.397985] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "17f34d32-e690-40d8-99a4-acf53fa859c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.398878] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.642359] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106896, 'name': CreateVM_Task, 'duration_secs': 0.407258} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.642559] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 954.643090] env[65121]: WARNING neutronclient.v2_0.client [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.643473] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.643619] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.644060] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 954.644338] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17788997-3e3c-42d2-ab59-0eca2ca448ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.649888] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 954.649888] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52159a0f-db02-abc2-ff31-e57da584e3dc" [ 954.649888] env[65121]: _type = "Task" [ 954.649888] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.659942] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52159a0f-db02-abc2-ff31-e57da584e3dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.702793] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5249a149-71db-2882-f56a-6a5c84aa8e8b, 'name': SearchDatastore_Task, 'duration_secs': 0.02365} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.703686] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e112ecd-e3ca-45ff-803f-801925a7aa6f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.709838] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 954.709838] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cccfb0-9233-0c0a-537e-6ac70aab22c2" [ 954.709838] env[65121]: _type = "Task" [ 954.709838] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.718669] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cccfb0-9233-0c0a-537e-6ac70aab22c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.733561] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "refresh_cache-68e3e4e6-6ca1-4f76-a4af-5112a2042c61" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.733898] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquired lock "refresh_cache-68e3e4e6-6ca1-4f76-a4af-5112a2042c61" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.734022] env[65121]: DEBUG nova.network.neutron [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 954.735089] env[65121]: DEBUG oslo_vmware.api [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5106898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207525} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.736103] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.736103] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.736103] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.765868] env[65121]: INFO nova.scheduler.client.report [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted allocations for instance e0d3e3c8-6471-4345-8677-369612674769 [ 954.782821] env[65121]: DEBUG nova.compute.utils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 954.785973] env[65121]: DEBUG nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 954.786275] env[65121]: DEBUG nova.network.neutron [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 954.786621] env[65121]: WARNING neutronclient.v2_0.client [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.786990] env[65121]: WARNING neutronclient.v2_0.client [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.787639] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.788039] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.831066] env[65121]: DEBUG nova.compute.manager [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 954.831660] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.832188] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c80b5a-7211-400e-a32f-82f382e4d830 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.843462] env[65121]: DEBUG nova.policy [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 954.848795] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.849262] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1205e21-098d-4cae-b74b-6da20c594b30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.860420] env[65121]: DEBUG oslo_vmware.api [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 954.860420] env[65121]: value = "task-5106899" [ 954.860420] env[65121]: _type = "Task" [ 954.860420] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.875983] env[65121]: DEBUG oslo_vmware.api [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106899, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.904646] env[65121]: DEBUG nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 954.965615] env[65121]: DEBUG nova.compute.manager [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Received event network-changed-86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 954.965615] env[65121]: DEBUG nova.compute.manager [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Refreshing instance network info cache due to event network-changed-86149f88-993f-45cf-94be-cb2781e74ac9. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 954.965615] env[65121]: DEBUG oslo_concurrency.lockutils [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Acquiring lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.965615] env[65121]: DEBUG oslo_concurrency.lockutils [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Acquired lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.965615] env[65121]: DEBUG nova.network.neutron [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Refreshing network info cache for port 86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 955.165097] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52159a0f-db02-abc2-ff31-e57da584e3dc, 'name': SearchDatastore_Task, 'duration_secs': 0.011308} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.166443] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.166724] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.166953] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.226838] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cccfb0-9233-0c0a-537e-6ac70aab22c2, 'name': SearchDatastore_Task, 'duration_secs': 0.011711} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.227232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.228016] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 6fe86deb-0903-4769-a05c-b7d0acec0103/6fe86deb-0903-4769-a05c-b7d0acec0103.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.228413] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 955.228619] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.228949] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3732616-9faa-47cb-b23d-f6678cb6a19d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.231217] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7cd61bf5-2247-4dad-a757-948c998b797a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.238456] env[65121]: WARNING neutronclient.v2_0.client [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.238937] env[65121]: WARNING openstack [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.239292] env[65121]: WARNING openstack [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.250837] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 955.250837] env[65121]: value = "task-5106900" [ 955.250837] env[65121]: _type = "Task" [ 955.250837] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.250972] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.252250] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.252250] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3779473-60e9-4ae5-a375-612a2e02eca9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.262639] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 955.262639] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527e14d8-3ae9-9455-2767-e9d86edf1055" [ 955.262639] env[65121]: _type = "Task" [ 955.262639] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.270155] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.270155] env[65121]: DEBUG nova.network.neutron [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 955.272467] env[65121]: DEBUG nova.network.neutron [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Successfully created port: f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 955.279802] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.288096] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527e14d8-3ae9-9455-2767-e9d86edf1055, 'name': SearchDatastore_Task, 'duration_secs': 0.011472} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.294918] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Received event network-changed-e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 955.295214] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Refreshing instance network info cache due to event network-changed-e31cbeae-05bc-416d-9df3-10f09d947ba4. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 955.295431] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquiring lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.295565] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquired lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 955.295755] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Refreshing network info cache for port e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 955.298314] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43055f0c-bbdb-4c00-a584-e01690394ee4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.301234] env[65121]: DEBUG nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 955.313336] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 955.313336] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e783f0-cc28-11d6-e34c-73fdd30285b1" [ 955.313336] env[65121]: _type = "Task" [ 955.313336] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.318291] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a22b127-11b5-4337-9662-d39376887ffe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.330903] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e783f0-cc28-11d6-e34c-73fdd30285b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.332258] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16125f3-ccb4-418b-8e6c-a0edd47125e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.375987] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482a0ab7-6769-47eb-8dbc-b92de2b61814 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.389945] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637b2cf1-482e-44ef-be4c-8e463f959d16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.394018] env[65121]: DEBUG oslo_vmware.api [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106899, 'name': PowerOffVM_Task, 'duration_secs': 0.281657} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.394693] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.396166] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.396166] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b65b4e7-6cb4-458a-9a1c-fe418a81bf33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.407825] env[65121]: DEBUG nova.compute.provider_tree [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.436026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.438910] env[65121]: DEBUG nova.network.neutron [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 955.472266] env[65121]: WARNING neutronclient.v2_0.client [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.472266] env[65121]: WARNING openstack [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.472266] env[65121]: WARNING openstack [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.478817] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.478817] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.478984] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleting the datastore file [datastore1] 59d6005d-6caf-4898-b791-70d9c015cdb4 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.480159] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e60e55d1-e9a6-4eb2-9e38-f8c0d70a3e81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.490271] env[65121]: DEBUG oslo_vmware.api [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 955.490271] env[65121]: value = "task-5106902" [ 955.490271] env[65121]: _type = "Task" [ 955.490271] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.501194] env[65121]: DEBUG oslo_vmware.api [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106902, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.582198] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "564159fe-6ecd-4276-8d65-cbe25859493a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.582579] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.606152] env[65121]: WARNING openstack [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.606939] env[65121]: WARNING openstack [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.727214] env[65121]: WARNING neutronclient.v2_0.client [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.727979] env[65121]: WARNING openstack [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.728386] env[65121]: WARNING openstack [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.765268] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106900, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.799480] env[65121]: WARNING neutronclient.v2_0.client [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.800373] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.800747] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.831591] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e783f0-cc28-11d6-e34c-73fdd30285b1, 'name': SearchDatastore_Task, 'duration_secs': 0.020901} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.831983] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.832443] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 59ceb0fd-1fb5-4c90-963d-fe76b9740d29/59ceb0fd-1fb5-4c90-963d-fe76b9740d29.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.832783] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f26d700-50a0-4d75-a6f5-8bc31747b222 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.843551] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 955.843551] env[65121]: value = "task-5106903" [ 955.843551] env[65121]: _type = "Task" [ 955.843551] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.856244] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106903, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.915884] env[65121]: DEBUG nova.network.neutron [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Updated VIF entry in instance network info cache for port 86149f88-993f-45cf-94be-cb2781e74ac9. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 955.916341] env[65121]: DEBUG nova.network.neutron [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Updating instance_info_cache with network_info: [{"id": "86149f88-993f-45cf-94be-cb2781e74ac9", "address": "fa:16:3e:e4:05:3a", "network": {"id": "9bd97a6f-a92f-49c4-a30f-c317253f98be", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1625344822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65e498022aee482da65a8176ab61fc71", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86149f88-99", "ovs_interfaceid": "86149f88-993f-45cf-94be-cb2781e74ac9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 955.919649] env[65121]: DEBUG nova.scheduler.client.report [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 955.942020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Releasing lock "refresh_cache-68e3e4e6-6ca1-4f76-a4af-5112a2042c61" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.942508] env[65121]: DEBUG nova.compute.manager [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 955.942738] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.944126] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335237f2-5cd4-4ced-825c-015c14ddbe40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.953438] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.958330] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4af051b4-e04f-4f48-adb1-20d344aa31e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.966778] env[65121]: DEBUG oslo_vmware.api [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 955.966778] env[65121]: value = "task-5106904" [ 955.966778] env[65121]: _type = "Task" [ 955.966778] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.978162] env[65121]: DEBUG oslo_vmware.api [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106904, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.002337] env[65121]: DEBUG oslo_vmware.api [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106902, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.005362] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.005362] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.085958] env[65121]: DEBUG nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 956.104058] env[65121]: WARNING neutronclient.v2_0.client [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.104250] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.104589] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.208546] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updated VIF entry in instance network info cache for port e31cbeae-05bc-416d-9df3-10f09d947ba4. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 956.208546] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating instance_info_cache with network_info: [{"id": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "address": "fa:16:3e:0e:bf:f3", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape31cbeae-05", "ovs_interfaceid": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.263845] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106900, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654052} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.264113] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 6fe86deb-0903-4769-a05c-b7d0acec0103/6fe86deb-0903-4769-a05c-b7d0acec0103.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.264273] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.264579] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c84ab11a-2f5b-4918-91d5-6503362b3b9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.273683] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 956.273683] env[65121]: value = "task-5106905" [ 956.273683] env[65121]: _type = "Task" [ 956.273683] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.285045] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106905, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.326555] env[65121]: DEBUG nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 956.357170] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106903, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.364542] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 956.364756] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 956.364906] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 956.368635] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 956.368635] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 956.368635] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 956.368887] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.368887] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 956.369132] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 956.369307] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 956.369497] env[65121]: DEBUG nova.virt.hardware [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 956.370815] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a948ce-8560-4a2f-ba3c-a3ad5c9e819c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.380920] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bcf4b9-ce21-4fb0-bcbf-a7b01686d6bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.422810] env[65121]: DEBUG oslo_concurrency.lockutils [req-53fe6eee-db18-411c-be32-6dbeec5cda2d req-fae2c9e6-782e-4e21-9a54-f65e9678b329 service nova] Releasing lock "refresh_cache-90cb98b1-9520-4caf-b6c4-80abaa7a9221" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 956.482398] env[65121]: DEBUG oslo_vmware.api [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106904, 'name': PowerOffVM_Task, 'duration_secs': 0.154395} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.482398] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.482398] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.482398] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e8a6e28-6d88-4e34-a9d3-93db8c62766f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.501737] env[65121]: DEBUG oslo_vmware.api [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106902, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.520193} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.502048] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.502228] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.502428] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.502649] env[65121]: INFO nova.compute.manager [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Took 1.67 seconds to destroy the instance on the hypervisor. [ 956.502857] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 956.503075] env[65121]: DEBUG nova.compute.manager [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 956.503172] env[65121]: DEBUG nova.network.neutron [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 956.503425] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.503966] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.504253] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.514364] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.514599] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.514778] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleting the datastore file [datastore2] 68e3e4e6-6ca1-4f76-a4af-5112a2042c61 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.515384] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a8e9239-ed63-45a2-aab1-8c150808869a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.524278] env[65121]: DEBUG oslo_vmware.api [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for the task: (returnval){ [ 956.524278] env[65121]: value = "task-5106907" [ 956.524278] env[65121]: _type = "Task" [ 956.524278] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.536647] env[65121]: DEBUG oslo_vmware.api [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.567533] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.616694] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.710188] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Releasing lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 956.710462] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Received event network-vif-plugged-ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 956.710653] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquiring lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.710852] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.711017] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.711181] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] No waiting events found dispatching network-vif-plugged-ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 956.711338] env[65121]: WARNING nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Received unexpected event network-vif-plugged-ad549932-0308-4165-96b1-9b4e419b6d05 for instance with vm_state building and task_state spawning. [ 956.711501] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Received event network-changed-ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 956.712639] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Refreshing instance network info cache due to event network-changed-ad549932-0308-4165-96b1-9b4e419b6d05. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 956.712639] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquiring lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.712639] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquired lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 956.712639] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Refreshing network info cache for port ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 956.786884] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106905, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.379632} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.786884] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.786884] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5eee5a-4bdf-483d-8f81-9c90f9bf9555 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.809560] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 6fe86deb-0903-4769-a05c-b7d0acec0103/6fe86deb-0903-4769-a05c-b7d0acec0103.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.809857] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-391c6bcc-f7fd-4f2c-8b4f-2f27237a6237 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.833102] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 956.833102] env[65121]: value = "task-5106908" [ 956.833102] env[65121]: _type = "Task" [ 956.833102] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.844875] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106908, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.856679] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106903, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.782929} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.857055] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 59ceb0fd-1fb5-4c90-963d-fe76b9740d29/59ceb0fd-1fb5-4c90-963d-fe76b9740d29.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.857321] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.857627] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e259b226-1f67-4c7d-921c-91817455d6ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.866829] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 956.866829] env[65121]: value = "task-5106909" [ 956.866829] env[65121]: _type = "Task" [ 956.866829] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.877191] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.922360] env[65121]: DEBUG nova.network.neutron [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Successfully updated port: f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 956.929602] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.653s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.929943] env[65121]: DEBUG nova.compute.manager [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=65121) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5433}} [ 956.933402] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.764s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.933713] env[65121]: DEBUG nova.objects.instance [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lazy-loading 'resources' on Instance uuid c52a1269-bb34-4ef6-ab8d-78df3e9b1d39 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.037409] env[65121]: DEBUG oslo_vmware.api [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Task: {'id': task-5106907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389852} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.037409] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.037631] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.037790] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.037982] env[65121]: INFO nova.compute.manager [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Took 1.10 seconds to destroy the instance on the hypervisor. [ 957.038349] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 957.038636] env[65121]: DEBUG nova.compute.manager [-] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 957.038784] env[65121]: DEBUG nova.network.neutron [-] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 957.039082] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.039789] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.040121] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.071529] env[65121]: DEBUG nova.network.neutron [-] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 957.072093] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.218369] env[65121]: WARNING neutronclient.v2_0.client [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.218871] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.219356] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.344184] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106908, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.347047] env[65121]: DEBUG nova.network.neutron [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.355268] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.355577] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.367094] env[65121]: DEBUG nova.compute.manager [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-vif-plugged-f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 957.367247] env[65121]: DEBUG oslo_concurrency.lockutils [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.367446] env[65121]: DEBUG oslo_concurrency.lockutils [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.367604] env[65121]: DEBUG oslo_concurrency.lockutils [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.367756] env[65121]: DEBUG nova.compute.manager [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] No waiting events found dispatching network-vif-plugged-f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 957.367908] env[65121]: WARNING nova.compute.manager [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received unexpected event network-vif-plugged-f696eb6b-11cc-4a52-9742-e5de719ed10b for instance with vm_state building and task_state spawning. [ 957.368080] env[65121]: DEBUG nova.compute.manager [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-changed-f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 957.368216] env[65121]: DEBUG nova.compute.manager [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing instance network info cache due to event network-changed-f696eb6b-11cc-4a52-9742-e5de719ed10b. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 957.368381] env[65121]: DEBUG oslo_concurrency.lockutils [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.368511] env[65121]: DEBUG oslo_concurrency.lockutils [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.368809] env[65121]: DEBUG nova.network.neutron [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing network info cache for port f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 957.380675] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188216} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.381107] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.381896] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8fb0b1-d88e-4b13-8f00-b0d27f29b4bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.416690] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 59ceb0fd-1fb5-4c90-963d-fe76b9740d29/59ceb0fd-1fb5-4c90-963d-fe76b9740d29.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.416690] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38d22f8e-f103-4680-aa00-d6a8110f09fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.433971] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.445316] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 957.445316] env[65121]: value = "task-5106910" [ 957.445316] env[65121]: _type = "Task" [ 957.445316] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.445893] env[65121]: WARNING neutronclient.v2_0.client [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.446513] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.446894] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.467687] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106910, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.515129] env[65121]: INFO nova.scheduler.client.report [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted allocation for migration 71f1858e-40ac-4394-b0fc-2d77e0195ba7 [ 957.557450] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Updated VIF entry in instance network info cache for port ad549932-0308-4165-96b1-9b4e419b6d05. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 957.557821] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Updating instance_info_cache with network_info: [{"id": "ad549932-0308-4165-96b1-9b4e419b6d05", "address": "fa:16:3e:78:7d:b1", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad549932-03", "ovs_interfaceid": "ad549932-0308-4165-96b1-9b4e419b6d05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.579283] env[65121]: DEBUG nova.network.neutron [-] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.601034] env[65121]: DEBUG nova.compute.manager [req-e5eb1701-7a3a-4947-8214-17574b66daaf req-248b2546-6c62-4a02-b5e9-3ab997dcce95 service nova] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Received event network-vif-deleted-920563c9-c663-4389-9348-0d9ccbca5fea {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 957.845635] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106908, 'name': ReconfigVM_Task, 'duration_secs': 0.872445} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.845948] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 6fe86deb-0903-4769-a05c-b7d0acec0103/6fe86deb-0903-4769-a05c-b7d0acec0103.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.847522] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-727cf597-5b0d-4663-aa8c-24a47cad5797 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.849268] env[65121]: INFO nova.compute.manager [-] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Took 1.35 seconds to deallocate network for instance. [ 957.858707] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 957.858707] env[65121]: value = "task-5106911" [ 957.858707] env[65121]: _type = "Task" [ 957.858707] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.870851] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106911, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.873975] env[65121]: WARNING neutronclient.v2_0.client [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.874615] env[65121]: WARNING openstack [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.874956] env[65121]: WARNING openstack [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.921296] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2292615-4db7-4744-a3a3-4487485e65a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.925223] env[65121]: DEBUG nova.network.neutron [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 957.933404] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25778c0-3453-4692-9f00-4cd4c3917ed5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.972293] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd4426c-3827-4e3d-83af-3d6a21365504 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.982107] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106910, 'name': ReconfigVM_Task, 'duration_secs': 0.293398} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.984764] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 59ceb0fd-1fb5-4c90-963d-fe76b9740d29/59ceb0fd-1fb5-4c90-963d-fe76b9740d29.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.985778] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64838eff-11b1-4491-8b3b-5fc4bace3eaf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.988497] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a84981-b897-4513-b286-c888d6e20439 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.003479] env[65121]: DEBUG nova.compute.provider_tree [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.006573] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 958.006573] env[65121]: value = "task-5106912" [ 958.006573] env[65121]: _type = "Task" [ 958.006573] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.023452] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8510b6f8-4a70-4e8b-a7e1-34c22dd3b845 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 31.595s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.033035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.061225] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Releasing lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.061554] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-vif-unplugged-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 958.061800] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.062069] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 958.062275] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.062479] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] No waiting events found dispatching network-vif-unplugged-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 958.062687] env[65121]: WARNING nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received unexpected event network-vif-unplugged-71e0942f-5026-4128-ba81-16311feb9b3e for instance with vm_state shelved_offloaded and task_state None. [ 958.062889] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 958.063095] env[65121]: DEBUG nova.compute.manager [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing instance network info cache due to event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 958.063319] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquiring lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.063487] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Acquired lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.063846] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 958.081889] env[65121]: INFO nova.compute.manager [-] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Took 1.04 seconds to deallocate network for instance. [ 958.095123] env[65121]: DEBUG nova.network.neutron [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 958.356676] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.367445] env[65121]: DEBUG nova.objects.instance [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'flavor' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.372872] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106911, 'name': Rename_Task, 'duration_secs': 0.388792} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.373531] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.373866] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d04959c-285e-4419-b25b-ace23586bfc3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.380760] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 958.380760] env[65121]: value = "task-5106913" [ 958.380760] env[65121]: _type = "Task" [ 958.380760] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.390244] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.511111] env[65121]: DEBUG nova.scheduler.client.report [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 958.526370] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106912, 'name': Rename_Task, 'duration_secs': 0.163307} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.526699] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.526990] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78e7070d-cca3-4545-aac2-94379b87db41 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.535085] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 958.535085] env[65121]: value = "task-5106914" [ 958.535085] env[65121]: _type = "Task" [ 958.535085] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.550964] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.567408] env[65121]: WARNING neutronclient.v2_0.client [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.568214] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.568656] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.590591] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.597558] env[65121]: DEBUG oslo_concurrency.lockutils [req-4e6eedbe-f005-4370-88f8-0e6e8f31b8b3 req-1f7b1cab-df35-4f94-b9e4-34adcd0bf7fe service nova] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.597967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.598815] env[65121]: DEBUG nova.network.neutron [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 958.716373] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.716934] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.794441] env[65121]: WARNING neutronclient.v2_0.client [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.795214] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.795560] env[65121]: WARNING openstack [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.876845] env[65121]: DEBUG oslo_concurrency.lockutils [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.877104] env[65121]: DEBUG oslo_concurrency.lockutils [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.877421] env[65121]: DEBUG nova.network.neutron [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 958.877632] env[65121]: DEBUG nova.objects.instance [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'info_cache' on Instance uuid 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.892361] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106913, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.937142] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updated VIF entry in instance network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 958.937565] env[65121]: DEBUG nova.network.neutron [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap71e0942f-50", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 959.020545] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.084s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.025124] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.835s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.025589] env[65121]: DEBUG nova.objects.instance [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 959.049121] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106914, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.056347] env[65121]: INFO nova.scheduler.client.report [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted allocations for instance c52a1269-bb34-4ef6-ab8d-78df3e9b1d39 [ 959.101254] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.101693] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.157464] env[65121]: DEBUG nova.network.neutron [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 959.180103] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.181136] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.267232] env[65121]: WARNING neutronclient.v2_0.client [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.268028] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.268469] env[65121]: WARNING openstack [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.368260] env[65121]: DEBUG nova.network.neutron [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 959.381313] env[65121]: DEBUG nova.objects.base [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Object Instance<7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c> lazy-loaded attributes: flavor,info_cache {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 959.396155] env[65121]: DEBUG oslo_vmware.api [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5106913, 'name': PowerOnVM_Task, 'duration_secs': 0.789518} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.396597] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.397220] env[65121]: INFO nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Took 10.22 seconds to spawn the instance on the hypervisor. [ 959.397583] env[65121]: DEBUG nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 959.399711] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20eca8a-aca8-4073-8900-793a14c4105b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.440426] env[65121]: DEBUG oslo_concurrency.lockutils [req-0a1da316-2d85-4ce4-9fd7-0cd929e91217 req-bee23f39-f196-4c8e-afc6-ff14818d91fb service nova] Releasing lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.548402] env[65121]: DEBUG oslo_vmware.api [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5106914, 'name': PowerOnVM_Task, 'duration_secs': 0.532926} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.548687] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.548881] env[65121]: INFO nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Took 8.06 seconds to spawn the instance on the hypervisor. [ 959.549060] env[65121]: DEBUG nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 959.549868] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b99509-d0c3-4b36-a12f-0865671ecc67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.563880] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf656a83-c771-42fb-ae2e-78f5794aa5ef tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "c52a1269-bb34-4ef6-ab8d-78df3e9b1d39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.605s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.873430] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.874060] env[65121]: DEBUG nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Instance network_info: |[{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 959.874624] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:b9:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9ec24851-7bb6-426b-b28f-f7b246df1713', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f696eb6b-11cc-4a52-9742-e5de719ed10b', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.883457] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 959.884204] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.884505] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dba8162e-1c85-4cb0-8a08-1b5efdf42f7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.905111] env[65121]: WARNING neutronclient.v2_0.client [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.905888] env[65121]: WARNING openstack [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.906259] env[65121]: WARNING openstack [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.920303] env[65121]: INFO nova.compute.manager [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Took 40.63 seconds to build instance. [ 959.923886] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.923886] env[65121]: value = "task-5106915" [ 959.923886] env[65121]: _type = "Task" [ 959.923886] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.936069] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106915, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.037524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8aae5859-d1e3-4fae-aeb6-0e6469d3072d tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.038993] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.028s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.039306] env[65121]: DEBUG nova.objects.instance [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'resources' on Instance uuid df5abcb5-583f-4b28-a074-3a3221d74d87 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.062194] env[65121]: WARNING openstack [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.062809] env[65121]: WARNING openstack [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.080525] env[65121]: INFO nova.compute.manager [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Took 37.81 seconds to build instance. [ 960.144846] env[65121]: WARNING neutronclient.v2_0.client [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.145582] env[65121]: WARNING openstack [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.145972] env[65121]: WARNING openstack [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.296322] env[65121]: DEBUG nova.network.neutron [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [{"id": "0e25d678-e865-4027-804f-77e9bb4cb31a", "address": "fa:16:3e:86:7e:5a", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e25d678-e8", "ovs_interfaceid": "0e25d678-e865-4027-804f-77e9bb4cb31a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 960.423324] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7cd83cb6-e82a-4a42-8dac-eade4ce13241 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.144s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.443024] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106915, 'name': CreateVM_Task, 'duration_secs': 0.433986} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.443024] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.443024] env[65121]: WARNING neutronclient.v2_0.client [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.443024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.443024] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 960.448852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 960.448852] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce04c673-0135-43aa-a6be-aec05888bffb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.450343] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 960.450343] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bf142a-82c6-92fa-d7aa-859980f3520a" [ 960.450343] env[65121]: _type = "Task" [ 960.450343] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.461733] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bf142a-82c6-92fa-d7aa-859980f3520a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.540907] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "efb98d37-4162-4249-9f85-008d4537db87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 960.541277] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "efb98d37-4162-4249-9f85-008d4537db87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.583275] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d52627f5-bcee-46c6-85b3-0ba661b38e1b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.327s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.809766] env[65121]: DEBUG oslo_concurrency.lockutils [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 960.965489] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bf142a-82c6-92fa-d7aa-859980f3520a, 'name': SearchDatastore_Task, 'duration_secs': 0.012276} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.966679] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 960.966679] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.966679] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.966679] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 960.966679] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.967124] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe0977dc-ca8a-4774-86ec-872cad3a78be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.980869] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.981084] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.985844] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-355bf766-8b2e-4e08-b503-c6044f828303 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.993306] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 960.993306] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522e5458-92f8-05ee-fc23-0e896d41479b" [ 960.993306] env[65121]: _type = "Task" [ 960.993306] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.002960] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522e5458-92f8-05ee-fc23-0e896d41479b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.007271] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883910e3-868b-441e-b965-e092a4ae72fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.016115] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260b348d-9240-4065-8abe-78e52a882d53 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.058453] env[65121]: DEBUG nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 961.067148] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4448d1-560b-477b-8595-3ca69e70b46e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.076249] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204eac5e-3314-425a-958d-4c009a47dbcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.094279] env[65121]: DEBUG nova.compute.provider_tree [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.185322] env[65121]: DEBUG nova.compute.manager [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Received event network-changed-e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 961.185543] env[65121]: DEBUG nova.compute.manager [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Refreshing instance network info cache due to event network-changed-e31cbeae-05bc-416d-9df3-10f09d947ba4. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 961.185772] env[65121]: DEBUG oslo_concurrency.lockutils [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Acquiring lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.185916] env[65121]: DEBUG oslo_concurrency.lockutils [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Acquired lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.187063] env[65121]: DEBUG nova.network.neutron [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Refreshing network info cache for port e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 961.443675] env[65121]: DEBUG nova.compute.manager [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Received event network-changed-ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 961.443675] env[65121]: DEBUG nova.compute.manager [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Refreshing instance network info cache due to event network-changed-ad549932-0308-4165-96b1-9b4e419b6d05. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 961.444082] env[65121]: DEBUG oslo_concurrency.lockutils [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Acquiring lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.444082] env[65121]: DEBUG oslo_concurrency.lockutils [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Acquired lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.444170] env[65121]: DEBUG nova.network.neutron [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Refreshing network info cache for port ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 961.504117] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522e5458-92f8-05ee-fc23-0e896d41479b, 'name': SearchDatastore_Task, 'duration_secs': 0.011008} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.504935] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d98cba8-6ede-4566-ae00-49921e9c4af9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.512723] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 961.512723] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52532dfc-f7c7-61a8-1857-41df97d452e3" [ 961.512723] env[65121]: _type = "Task" [ 961.512723] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.523139] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52532dfc-f7c7-61a8-1857-41df97d452e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.598614] env[65121]: DEBUG nova.scheduler.client.report [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 961.686793] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.692609] env[65121]: WARNING neutronclient.v2_0.client [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.692609] env[65121]: WARNING openstack [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.692609] env[65121]: WARNING openstack [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.821455] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.821796] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-061618a4-386f-4b9d-9a8b-339fa8302955 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.829983] env[65121]: WARNING openstack [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.830507] env[65121]: WARNING openstack [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.839242] env[65121]: DEBUG oslo_vmware.api [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 961.839242] env[65121]: value = "task-5106916" [ 961.839242] env[65121]: _type = "Task" [ 961.839242] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.851604] env[65121]: DEBUG oslo_vmware.api [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106916, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.907731] env[65121]: WARNING neutronclient.v2_0.client [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.907926] env[65121]: WARNING openstack [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.908271] env[65121]: WARNING openstack [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.947034] env[65121]: WARNING neutronclient.v2_0.client [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.947597] env[65121]: WARNING openstack [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.947979] env[65121]: WARNING openstack [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 962.000615] env[65121]: DEBUG nova.network.neutron [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updated VIF entry in instance network info cache for port e31cbeae-05bc-416d-9df3-10f09d947ba4. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 962.001168] env[65121]: DEBUG nova.network.neutron [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating instance_info_cache with network_info: [{"id": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "address": "fa:16:3e:0e:bf:f3", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape31cbeae-05", "ovs_interfaceid": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 962.026287] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52532dfc-f7c7-61a8-1857-41df97d452e3, 'name': SearchDatastore_Task, 'duration_secs': 0.013823} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.026489] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.026716] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 23d6f632-8918-46af-b239-08a9615dfbec/23d6f632-8918-46af-b239-08a9615dfbec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.027096] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76690e08-37e2-42c5-ab83-225a90c8e98f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.042080] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 962.042080] env[65121]: value = "task-5106917" [ 962.042080] env[65121]: _type = "Task" [ 962.042080] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.053399] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106917, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.077971] env[65121]: WARNING openstack [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 962.078455] env[65121]: WARNING openstack [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 962.105915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.067s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 962.113593] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.042s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 962.113593] env[65121]: DEBUG nova.objects.instance [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lazy-loading 'resources' on Instance uuid 7fcab9cb-9a0f-4a50-9244-687889443a31 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.135498] env[65121]: INFO nova.scheduler.client.report [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted allocations for instance df5abcb5-583f-4b28-a074-3a3221d74d87 [ 962.157776] env[65121]: WARNING neutronclient.v2_0.client [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 962.158909] env[65121]: WARNING openstack [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 962.159338] env[65121]: WARNING openstack [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 962.273148] env[65121]: DEBUG nova.network.neutron [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Updated VIF entry in instance network info cache for port ad549932-0308-4165-96b1-9b4e419b6d05. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 962.273566] env[65121]: DEBUG nova.network.neutron [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Updating instance_info_cache with network_info: [{"id": "ad549932-0308-4165-96b1-9b4e419b6d05", "address": "fa:16:3e:78:7d:b1", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad549932-03", "ovs_interfaceid": "ad549932-0308-4165-96b1-9b4e419b6d05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 962.351327] env[65121]: DEBUG oslo_vmware.api [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106916, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.504057] env[65121]: DEBUG oslo_concurrency.lockutils [req-21841bb7-3c7b-4f42-8c3b-2b6cf9342d4b req-d2fb59d9-5b21-4d1a-8d5d-4bee04c7550d service nova] Releasing lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.553181] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106917, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.644629] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2be63251-0749-4ccc-8f2e-796466afa00d tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "df5abcb5-583f-4b28-a074-3a3221d74d87" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.411s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 962.776567] env[65121]: DEBUG oslo_concurrency.lockutils [req-611216db-a9ba-43c9-963e-394d2a72b7a8 req-aff0a467-2cc4-4fe8-b561-a89e938e81bf service nova] Releasing lock "refresh_cache-59ceb0fd-1fb5-4c90-963d-fe76b9740d29" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.853272] env[65121]: DEBUG oslo_vmware.api [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106916, 'name': PowerOnVM_Task, 'duration_secs': 0.588246} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.853874] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.853874] env[65121]: DEBUG nova.compute.manager [None req-afd0ee0c-1a2d-4a56-9adb-e9da04ed6716 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 962.854735] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a9690f-0e66-43e5-83cf-49fd38995a45 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.053567] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106917, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.766156} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.056615] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 23d6f632-8918-46af-b239-08a9615dfbec/23d6f632-8918-46af-b239-08a9615dfbec.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.056845] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.057357] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-785901c9-ae12-42ad-988a-ce927009b096 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.064078] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 963.064078] env[65121]: value = "task-5106918" [ 963.064078] env[65121]: _type = "Task" [ 963.064078] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.077021] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106918, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.155596] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50daf4de-cecc-44bb-a1a2-62a93590d8ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.164604] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4303d1b7-f1ca-423a-bc53-d63e7d0c45c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.199832] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9374cb-24f1-4814-b424-05d134899b9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.210741] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff400f96-d39e-4e1e-93cb-c4a2d1e159ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.225814] env[65121]: DEBUG nova.compute.provider_tree [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.576605] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106918, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188623} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.576967] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.577798] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e57fbe-9452-4436-ad1c-87d12c60bfbf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.602196] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 23d6f632-8918-46af-b239-08a9615dfbec/23d6f632-8918-46af-b239-08a9615dfbec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.602534] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24ac1976-9e1d-48f0-8dbf-048e4438396a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.623399] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 963.623399] env[65121]: value = "task-5106919" [ 963.623399] env[65121]: _type = "Task" [ 963.623399] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.632784] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106919, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.732008] env[65121]: DEBUG nova.scheduler.client.report [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 964.136354] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106919, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.238063] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.241177] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.783s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.242179] env[65121]: DEBUG nova.objects.instance [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lazy-loading 'resources' on Instance uuid 2e676b64-181e-4fee-8120-05cb49b5bb4e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.262200] env[65121]: INFO nova.scheduler.client.report [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Deleted allocations for instance 7fcab9cb-9a0f-4a50-9244-687889443a31 [ 964.639782] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106919, 'name': ReconfigVM_Task, 'duration_secs': 0.898545} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.640125] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 23d6f632-8918-46af-b239-08a9615dfbec/23d6f632-8918-46af-b239-08a9615dfbec.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.640700] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-523bf116-4900-4cd4-afe0-293c52bf49be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.648430] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.648657] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.648854] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.649032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.649192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.651767] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 964.651767] env[65121]: value = "task-5106920" [ 964.651767] env[65121]: _type = "Task" [ 964.651767] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.655505] env[65121]: INFO nova.compute.manager [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Terminating instance [ 964.664923] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106920, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.770747] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c6b95a79-fd6d-4ca1-b310-96b601df1dda tempest-ServersTestJSON-646408521 tempest-ServersTestJSON-646408521-project-member] Lock "7fcab9cb-9a0f-4a50-9244-687889443a31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.489s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.161414] env[65121]: DEBUG nova.compute.manager [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 965.161414] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.161414] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49222fb-9b05-48c0-bb11-dd7705fc61b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.168327] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106920, 'name': Rename_Task, 'duration_secs': 0.218585} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.168754] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.171603] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc89ddcd-688d-4920-aa6c-27c44a6d8841 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.175666] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.176438] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4abecf4c-5c6f-4d30-b260-27dc108678db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.182569] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 965.182569] env[65121]: value = "task-5106921" [ 965.182569] env[65121]: _type = "Task" [ 965.182569] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.189716] env[65121]: DEBUG oslo_vmware.api [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 965.189716] env[65121]: value = "task-5106922" [ 965.189716] env[65121]: _type = "Task" [ 965.189716] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.197635] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106921, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.203522] env[65121]: DEBUG oslo_vmware.api [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106922, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.205435] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf01e29d-2159-4c8e-aa6b-0b433049dab2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.213137] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "6e969388-3238-404d-a8eb-e7b7318c4c72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.213389] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.218499] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1ea6ab-6d7c-4a61-9b24-b839b6c62a9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.254411] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb43bcd-2b35-4a5d-862e-f694b84fe6d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.264198] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6945a50a-0545-44d2-b163-ae664bcd2202 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.281374] env[65121]: DEBUG nova.compute.provider_tree [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.693309] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106921, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.703013] env[65121]: DEBUG oslo_vmware.api [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106922, 'name': PowerOffVM_Task, 'duration_secs': 0.26092} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.703432] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.703644] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.703945] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-481385ed-52eb-4d6f-a74f-dc9437231f14 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.718477] env[65121]: DEBUG nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 965.769856] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.770122] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.770322] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleting the datastore file [datastore1] 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.770574] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-806621ca-e9ec-4c6e-adc6-e80ffe374c16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.779650] env[65121]: DEBUG oslo_vmware.api [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 965.779650] env[65121]: value = "task-5106924" [ 965.779650] env[65121]: _type = "Task" [ 965.779650] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.784053] env[65121]: DEBUG nova.scheduler.client.report [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 965.791393] env[65121]: DEBUG oslo_vmware.api [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106924, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.196488] env[65121]: DEBUG oslo_vmware.api [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5106921, 'name': PowerOnVM_Task, 'duration_secs': 0.833668} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.196613] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.196895] env[65121]: INFO nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Took 9.87 seconds to spawn the instance on the hypervisor. [ 966.197105] env[65121]: DEBUG nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 966.197911] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2f2e1d-8532-4f16-83ae-b9f6fa5aa003 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.247758] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.290742] env[65121]: DEBUG oslo_vmware.api [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5106924, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248159} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.291137] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.291392] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.291590] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.291776] env[65121]: INFO nova.compute.manager [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 966.292098] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 966.292926] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.052s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.294922] env[65121]: DEBUG nova.compute.manager [-] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 966.295039] env[65121]: DEBUG nova.network.neutron [-] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 966.295277] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 966.295883] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 966.296516] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 966.304034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.626s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.304034] env[65121]: DEBUG nova.objects.instance [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 966.330328] env[65121]: INFO nova.scheduler.client.report [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Deleted allocations for instance 2e676b64-181e-4fee-8120-05cb49b5bb4e [ 966.342636] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 966.721514] env[65121]: INFO nova.compute.manager [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Took 38.51 seconds to build instance. [ 966.839687] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb23716e-25c6-40de-b0cd-6d2040607e62 tempest-ServersListShow2100Test-350046337 tempest-ServersListShow2100Test-350046337-project-member] Lock "2e676b64-181e-4fee-8120-05cb49b5bb4e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.727s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.958702] env[65121]: DEBUG nova.compute.manager [req-e82deea8-3c7b-400c-9dca-b90e7de5c4f4 req-98bb46e8-3c09-41b1-9c0f-d6020a8e7b99 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Received event network-vif-deleted-0e25d678-e865-4027-804f-77e9bb4cb31a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 966.958983] env[65121]: INFO nova.compute.manager [req-e82deea8-3c7b-400c-9dca-b90e7de5c4f4 req-98bb46e8-3c09-41b1-9c0f-d6020a8e7b99 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Neutron deleted interface 0e25d678-e865-4027-804f-77e9bb4cb31a; detaching it from the instance and deleting it from the info cache [ 966.959241] env[65121]: DEBUG nova.network.neutron [req-e82deea8-3c7b-400c-9dca-b90e7de5c4f4 req-98bb46e8-3c09-41b1-9c0f-d6020a8e7b99 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 967.225354] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a4e70270-f85f-40eb-976b-f1402823db7d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "23d6f632-8918-46af-b239-08a9615dfbec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.023s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.306347] env[65121]: DEBUG nova.network.neutron [-] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 967.320515] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6f81eee-b829-487c-ba74-b0d581a17d5c tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.322583] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.606s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.322583] env[65121]: DEBUG nova.objects.instance [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lazy-loading 'resources' on Instance uuid b2ebce1d-a8d8-4a71-997b-39d09ca373fe {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.462516] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bbf26357-4ca2-469e-8eca-2639992847d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.475281] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffacd05-9159-4fa7-a920-01b0632f2f2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.516776] env[65121]: DEBUG nova.compute.manager [req-e82deea8-3c7b-400c-9dca-b90e7de5c4f4 req-98bb46e8-3c09-41b1-9c0f-d6020a8e7b99 service nova] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Detach interface failed, port_id=0e25d678-e865-4027-804f-77e9bb4cb31a, reason: Instance 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 967.808223] env[65121]: INFO nova.compute.manager [-] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Took 1.51 seconds to deallocate network for instance. [ 968.277881] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebe9204-c73c-4f90-9e01-c9fd00f3213f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.286318] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeeacaf4-84d1-4fe5-8b0a-81e68f9fa458 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.320409] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.321503] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b94c914-344a-47b5-abd1-f03bebba709f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.330715] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e7d123-bc42-4e1e-9f95-7c5693d91ca6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.347595] env[65121]: DEBUG nova.compute.provider_tree [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.495039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 968.495287] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 968.852050] env[65121]: DEBUG nova.scheduler.client.report [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 968.998200] env[65121]: DEBUG nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 969.021984] env[65121]: DEBUG nova.compute.manager [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-changed-f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 969.022474] env[65121]: DEBUG nova.compute.manager [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing instance network info cache due to event network-changed-f696eb6b-11cc-4a52-9742-e5de719ed10b. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 969.023987] env[65121]: DEBUG oslo_concurrency.lockutils [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.023987] env[65121]: DEBUG oslo_concurrency.lockutils [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 969.024278] env[65121]: DEBUG nova.network.neutron [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing network info cache for port f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 969.357912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.038s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.361514] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.556s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 969.361926] env[65121]: DEBUG nova.objects.instance [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lazy-loading 'resources' on Instance uuid ab41ae41-e69d-47fb-a31a-16748fd3b0f2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.380080] env[65121]: INFO nova.scheduler.client.report [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted allocations for instance b2ebce1d-a8d8-4a71-997b-39d09ca373fe [ 969.518880] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 969.526909] env[65121]: WARNING neutronclient.v2_0.client [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.527620] env[65121]: WARNING openstack [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.527957] env[65121]: WARNING openstack [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.656249] env[65121]: WARNING openstack [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.656635] env[65121]: WARNING openstack [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.723934] env[65121]: WARNING neutronclient.v2_0.client [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.724740] env[65121]: WARNING openstack [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.725104] env[65121]: WARNING openstack [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.812170] env[65121]: DEBUG nova.network.neutron [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updated VIF entry in instance network info cache for port f696eb6b-11cc-4a52-9742-e5de719ed10b. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 969.812170] env[65121]: DEBUG nova.network.neutron [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 969.889444] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ae08486e-9fb5-4db6-a927-132f196924a9 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "b2ebce1d-a8d8-4a71-997b-39d09ca373fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.827s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.197849] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "3f3bec5b-2834-497c-a454-a152cb992309" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.197956] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "3f3bec5b-2834-497c-a454-a152cb992309" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.198581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "3f3bec5b-2834-497c-a454-a152cb992309-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.198581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "3f3bec5b-2834-497c-a454-a152cb992309-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.198581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "3f3bec5b-2834-497c-a454-a152cb992309-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.200958] env[65121]: INFO nova.compute.manager [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Terminating instance [ 970.260820] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b00b090-61ec-4c18-a11a-46156cbe1575 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.268418] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f48b38-c14e-4638-b552-0a97e6ed667e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.302800] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b66b3c-969c-44c5-91dc-c96b21df38bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.311789] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e411f72b-73df-4dc2-b1a7-4945f222a271 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.316228] env[65121]: DEBUG oslo_concurrency.lockutils [req-8b026072-0f92-4a84-aae7-54a1193b1e43 req-1ff2781c-a8ef-4694-84b7-9e59eba2ceb4 service nova] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 970.327330] env[65121]: DEBUG nova.compute.provider_tree [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.704848] env[65121]: DEBUG nova.compute.manager [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 970.705094] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.706293] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ccd2dd-ce62-45d2-b6e1-19b3e2ed4658 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.715420] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.715685] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c803b46f-4775-42da-93ee-8f44b835c478 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.721487] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "e756fdb2-f901-4e23-92e7-f2753fd5f728" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.721743] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.726380] env[65121]: DEBUG oslo_vmware.api [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 970.726380] env[65121]: value = "task-5106925" [ 970.726380] env[65121]: _type = "Task" [ 970.726380] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.736045] env[65121]: DEBUG oslo_vmware.api [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.830900] env[65121]: DEBUG nova.scheduler.client.report [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 971.224644] env[65121]: DEBUG nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 971.238747] env[65121]: DEBUG oslo_vmware.api [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106925, 'name': PowerOffVM_Task, 'duration_secs': 0.20198} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.238747] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.238942] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.239406] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d733812-75fa-4f90-a66c-43308187686c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.312107] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.312303] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.312483] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore1] 3f3bec5b-2834-497c-a454-a152cb992309 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.312753] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d373cc5-4d04-4725-9887-cbc8d05baee8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.320949] env[65121]: DEBUG oslo_vmware.api [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 971.320949] env[65121]: value = "task-5106927" [ 971.320949] env[65121]: _type = "Task" [ 971.320949] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.329794] env[65121]: DEBUG oslo_vmware.api [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.336509] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.975s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.339296] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.364s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.339296] env[65121]: DEBUG nova.objects.instance [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lazy-loading 'resources' on Instance uuid 49b7e6c3-9106-4b57-ae44-85b5d4474089 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.361451] env[65121]: INFO nova.scheduler.client.report [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Deleted allocations for instance ab41ae41-e69d-47fb-a31a-16748fd3b0f2 [ 971.748668] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.834604] env[65121]: DEBUG oslo_vmware.api [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147945} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.835033] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.835356] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.835650] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.835937] env[65121]: INFO nova.compute.manager [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Took 1.13 seconds to destroy the instance on the hypervisor. [ 971.836337] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 971.836614] env[65121]: DEBUG nova.compute.manager [-] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 971.836766] env[65121]: DEBUG nova.network.neutron [-] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 971.837234] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 971.838115] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.838546] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.870049] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f26cf980-248c-4230-a7e8-888fab3651e1 tempest-ServerRescueTestJSON-1357542986 tempest-ServerRescueTestJSON-1357542986-project-member] Lock "ab41ae41-e69d-47fb-a31a-16748fd3b0f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.066s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.881201] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.196949] env[65121]: DEBUG nova.compute.manager [req-533ff792-5151-403d-a19b-ea3e591be026 req-aaef4568-1e2e-4ad4-8259-0a4de386aa38 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Received event network-vif-deleted-30bb9dd1-838c-45e3-80f2-580a96b705ff {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 972.197379] env[65121]: INFO nova.compute.manager [req-533ff792-5151-403d-a19b-ea3e591be026 req-aaef4568-1e2e-4ad4-8259-0a4de386aa38 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Neutron deleted interface 30bb9dd1-838c-45e3-80f2-580a96b705ff; detaching it from the instance and deleting it from the info cache [ 972.197379] env[65121]: DEBUG nova.network.neutron [req-533ff792-5151-403d-a19b-ea3e591be026 req-aaef4568-1e2e-4ad4-8259-0a4de386aa38 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 972.294610] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f62ed0-2187-485c-ba5a-3c10aad8aeb7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.304089] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668a7963-aa15-4dad-96ac-1cb2d8615ac0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.336583] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6366f54-8125-498d-a1a5-8cb8672e9b84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.345215] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e008cd4-1fb4-4f79-a0d8-e63592981102 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.362516] env[65121]: DEBUG nova.compute.provider_tree [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.647980] env[65121]: DEBUG nova.network.neutron [-] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 972.702052] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc83e83f-c598-4716-89c7-89a55288fd4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.712139] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2ccdd0-ea73-47a6-a10e-9d5573d8cf00 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.749416] env[65121]: DEBUG nova.compute.manager [req-533ff792-5151-403d-a19b-ea3e591be026 req-aaef4568-1e2e-4ad4-8259-0a4de386aa38 service nova] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Detach interface failed, port_id=30bb9dd1-838c-45e3-80f2-580a96b705ff, reason: Instance 3f3bec5b-2834-497c-a454-a152cb992309 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 972.866341] env[65121]: DEBUG nova.scheduler.client.report [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 973.020625] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.020909] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.155801] env[65121]: INFO nova.compute.manager [-] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Took 1.32 seconds to deallocate network for instance. [ 973.372075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.033s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.375101] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.932s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.375101] env[65121]: DEBUG nova.objects.instance [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lazy-loading 'resources' on Instance uuid b6f6871d-311c-4adb-824e-2907a12f4224 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.399816] env[65121]: INFO nova.scheduler.client.report [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Deleted allocations for instance 49b7e6c3-9106-4b57-ae44-85b5d4474089 [ 973.531233] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.531233] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.531233] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.531233] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.531233] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.531233] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.531233] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 973.531233] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.662337] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.910492] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f35929ae-e804-4e7d-a24d-9cf13864735d tempest-ImagesOneServerNegativeTestJSON-1452745102 tempest-ImagesOneServerNegativeTestJSON-1452745102-project-member] Lock "49b7e6c3-9106-4b57-ae44-85b5d4474089" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.783s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.034244] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.306683] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3a3479-1851-45cb-9b66-69c2c54bcb08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.317386] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9a4d52-563a-48b5-9aeb-90f860aa6b9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.352051] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7fca15-1f99-4a04-a27a-57c8926801d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.360345] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc1417c-66bc-4b8d-a28c-f87977aaa09e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.380569] env[65121]: DEBUG nova.compute.provider_tree [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.883481] env[65121]: DEBUG nova.scheduler.client.report [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 975.389260] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.014s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.391957] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.538s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.392215] env[65121]: DEBUG nova.objects.instance [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'resources' on Instance uuid 600d5554-f52a-48ca-941d-1a755d086823 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.414452] env[65121]: INFO nova.scheduler.client.report [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted allocations for instance b6f6871d-311c-4adb-824e-2907a12f4224 [ 975.923604] env[65121]: DEBUG oslo_concurrency.lockutils [None req-12c43e56-c58c-4d4c-adfd-c4862f6bcd10 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "b6f6871d-311c-4adb-824e-2907a12f4224" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.041s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.360503] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75d9fd3-63c3-49ba-b2f1-931d62a0be9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.368631] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43842e38-991e-42a9-8439-17722cc97e1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.400337] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b697bab-2bf6-45eb-88bb-eabf07b76089 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.408436] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f41a4e7-50ff-4bc6-a972-a9e1abd632e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.422921] env[65121]: DEBUG nova.compute.provider_tree [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.643494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.643780] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 976.644034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.644194] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 976.644357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.647975] env[65121]: INFO nova.compute.manager [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Terminating instance [ 976.926875] env[65121]: DEBUG nova.scheduler.client.report [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.151758] env[65121]: DEBUG nova.compute.manager [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 977.151983] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 977.152937] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655b8ee7-438a-4222-958f-1c0063e94271 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.161533] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 977.161810] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-978eef11-3b77-4e90-8233-12ee411f9f12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.168705] env[65121]: DEBUG oslo_vmware.api [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 977.168705] env[65121]: value = "task-5106928" [ 977.168705] env[65121]: _type = "Task" [ 977.168705] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.179132] env[65121]: DEBUG oslo_vmware.api [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106928, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.438200] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.046s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.441441] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.275s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.441441] env[65121]: DEBUG nova.objects.instance [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lazy-loading 'resources' on Instance uuid ac001ee6-f601-48f7-af0d-42c2387f3c98 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.467413] env[65121]: INFO nova.scheduler.client.report [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted allocations for instance 600d5554-f52a-48ca-941d-1a755d086823 [ 977.680095] env[65121]: DEBUG oslo_vmware.api [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106928, 'name': PowerOffVM_Task, 'duration_secs': 0.256838} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.680404] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.680575] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 977.680844] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64d65fcd-fe11-487b-bcaa-499285e5ccab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.748649] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 977.748901] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 977.749105] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleting the datastore file [datastore1] 9cf6c29a-3424-4b88-9ba5-8120b124beb6 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.749380] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d367cda-55eb-43d9-bdd8-a9469483deaa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.756596] env[65121]: DEBUG oslo_vmware.api [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 977.756596] env[65121]: value = "task-5106930" [ 977.756596] env[65121]: _type = "Task" [ 977.756596] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.765997] env[65121]: DEBUG oslo_vmware.api [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106930, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.978773] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2d6b039-ab0a-41ee-b2cd-002025f951d9 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "600d5554-f52a-48ca-941d-1a755d086823" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.114s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.268388] env[65121]: DEBUG oslo_vmware.api [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106930, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145408} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.271288] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.271537] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.271719] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.271898] env[65121]: INFO nova.compute.manager [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 978.272214] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 978.272625] env[65121]: DEBUG nova.compute.manager [-] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 978.272721] env[65121]: DEBUG nova.network.neutron [-] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 978.272963] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 978.273722] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.274143] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.365259] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 978.407093] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6ddc76-a475-43c9-8490-66afa6db3529 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.429951] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277da6db-af48-4882-971a-bae96fba4ef5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.465601] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e7b943-cda9-446e-93c7-3c2a437a05a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.474291] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1103e790-afa6-43fb-ae72-5de04f6eca57 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.489870] env[65121]: DEBUG nova.compute.provider_tree [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.774965] env[65121]: DEBUG nova.compute.manager [req-23112856-ea6f-4b4f-918a-17d765728026 req-90f8913f-ae19-4832-a562-c47ef2e07cc1 service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Received event network-vif-deleted-df584c77-b7b3-4d0f-a9eb-6dece1793f87 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 978.775192] env[65121]: INFO nova.compute.manager [req-23112856-ea6f-4b4f-918a-17d765728026 req-90f8913f-ae19-4832-a562-c47ef2e07cc1 service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Neutron deleted interface df584c77-b7b3-4d0f-a9eb-6dece1793f87; detaching it from the instance and deleting it from the info cache [ 978.775362] env[65121]: DEBUG nova.network.neutron [req-23112856-ea6f-4b4f-918a-17d765728026 req-90f8913f-ae19-4832-a562-c47ef2e07cc1 service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 978.994263] env[65121]: DEBUG nova.scheduler.client.report [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.070464] env[65121]: DEBUG nova.network.neutron [-] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 979.278849] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32856bf0-0d11-439f-80d3-e18bc680b267 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.291559] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7525c5d0-1290-45b4-b5a7-77892d7321db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.328594] env[65121]: DEBUG nova.compute.manager [req-23112856-ea6f-4b4f-918a-17d765728026 req-90f8913f-ae19-4832-a562-c47ef2e07cc1 service nova] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Detach interface failed, port_id=df584c77-b7b3-4d0f-a9eb-6dece1793f87, reason: Instance 9cf6c29a-3424-4b88-9ba5-8120b124beb6 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 979.498580] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.500991] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.888s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.501340] env[65121]: DEBUG nova.objects.instance [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lazy-loading 'resources' on Instance uuid 1b6ad893-e013-4fd5-a829-535d1ea6e001 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.528477] env[65121]: INFO nova.scheduler.client.report [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Deleted allocations for instance ac001ee6-f601-48f7-af0d-42c2387f3c98 [ 979.572806] env[65121]: INFO nova.compute.manager [-] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Took 1.30 seconds to deallocate network for instance. [ 980.036297] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8869666c-8bcb-4b22-a8cc-6552108016b0 tempest-ServerMetadataTestJSON-1022422317 tempest-ServerMetadataTestJSON-1022422317-project-member] Lock "ac001ee6-f601-48f7-af0d-42c2387f3c98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.409s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 980.081533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.423711] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecca2b78-3bdb-4786-9b5f-67f8ca41bbfd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.433085] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602a990e-784b-42a3-895b-2732122b83e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.473058] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db75793-94cb-4864-b30b-155938a82c6d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.481199] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499466da-6e34-489e-b84b-b417b438b0e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.496073] env[65121]: DEBUG nova.compute.provider_tree [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.728440] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.728765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 980.775864] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "ec463cad-4c80-4636-bc7a-9ec298a07d96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.776189] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 980.810273] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.810522] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.000035] env[65121]: DEBUG nova.scheduler.client.report [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 981.150116] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.150638] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.231818] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 981.279781] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 981.312562] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 981.505565] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.004s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 981.508351] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.884s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.508650] env[65121]: DEBUG nova.objects.instance [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lazy-loading 'resources' on Instance uuid 5f6e4f46-0745-42c0-a779-6cffb60e21ca {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.528397] env[65121]: INFO nova.scheduler.client.report [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Deleted allocations for instance 1b6ad893-e013-4fd5-a829-535d1ea6e001 [ 981.755326] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.810753] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.839415] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.042168] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c831c586-c685-428c-a235-1b0acd2b1f50 tempest-ServersTestBootFromVolume-437173978 tempest-ServersTestBootFromVolume-437173978-project-member] Lock "1b6ad893-e013-4fd5-a829-535d1ea6e001" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.957s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.383865] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c045205a-75cb-4497-b340-00df71adbb7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.393451] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cad3ad7-7751-42aa-bd48-ce9b55afeafd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.429390] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8ec7eb-7a44-40c4-b745-0b9db29eab32 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.436860] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c39e417-ca6b-44ea-bbb1-4250de2f80ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.451352] env[65121]: DEBUG nova.compute.provider_tree [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.954302] env[65121]: DEBUG nova.scheduler.client.report [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 983.461876] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.953s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.464357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.530s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.464547] env[65121]: DEBUG nova.objects.instance [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 983.493557] env[65121]: INFO nova.scheduler.client.report [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Deleted allocations for instance 5f6e4f46-0745-42c0-a779-6cffb60e21ca [ 984.000998] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cb20d816-0b64-4f79-bb4f-60e9f7c8b7c3 tempest-MigrationsAdminTest-1477697275 tempest-MigrationsAdminTest-1477697275-project-member] Lock "5f6e4f46-0745-42c0-a779-6cffb60e21ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.560s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.480358] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0c105086-b9f6-4ae0-9a60-158212e49d6f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.481495] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.826s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.482969] env[65121]: INFO nova.compute.claims [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.933893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.934158] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 985.846065] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3752dc-6652-429f-880d-4b34d206f2ea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.854102] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728ca97e-54e0-4703-8b80-6deb2c8e62c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.887988] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6e8d4a-f23f-44f1-bbed-57006a4877bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.896439] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c34c649-d312-4802-bfab-e69759d03741 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.910510] env[65121]: DEBUG nova.compute.provider_tree [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.414818] env[65121]: DEBUG nova.scheduler.client.report [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 986.921605] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 986.921605] env[65121]: DEBUG nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 986.925235] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 34.233s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 986.925446] env[65121]: DEBUG nova.objects.instance [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 987.430161] env[65121]: DEBUG nova.compute.utils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 987.436073] env[65121]: DEBUG nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 987.436308] env[65121]: DEBUG nova.network.neutron [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 987.436685] env[65121]: WARNING neutronclient.v2_0.client [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.437021] env[65121]: WARNING neutronclient.v2_0.client [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.437756] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.438188] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.493124] env[65121]: DEBUG nova.policy [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd148efb6c7814df182782805ba0e326b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48809b7bf36c4315b9ac315d6b9ca0c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 987.792958] env[65121]: DEBUG nova.network.neutron [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Successfully created port: 24f646d7-31dd-416b-a619-d45d5e0fd1b2 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 987.937590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2b21325c-741a-465b-aed0-f67e6b9b1590 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.939955] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.660s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.939955] env[65121]: DEBUG nova.objects.instance [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'resources' on Instance uuid e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.948603] env[65121]: DEBUG nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 988.347407] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "11caa171-0314-4e7d-8502-80932b469da8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.347644] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "11caa171-0314-4e7d-8502-80932b469da8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.460406] env[65121]: DEBUG nova.objects.instance [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'numa_topology' on Instance uuid e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.965448] env[65121]: DEBUG nova.objects.base [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 988.969292] env[65121]: DEBUG nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 989.002448] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 989.002448] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 989.002742] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 989.002957] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 989.003172] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 989.003320] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 989.003625] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.003730] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 989.003922] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 989.004095] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 989.004287] env[65121]: DEBUG nova.virt.hardware [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 989.005736] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adfa008-3972-483c-83d3-e79efc8d6cf0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.017924] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be555c6e-8cd5-41f0-834a-82edfbfe81e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.345143] env[65121]: DEBUG nova.compute.manager [req-8a4d52b6-8071-4ca2-8ecd-72defceff925 req-b9638ce2-3ba9-4e61-8b7b-b2655fcf5058 service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Received event network-vif-plugged-24f646d7-31dd-416b-a619-d45d5e0fd1b2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 989.345389] env[65121]: DEBUG oslo_concurrency.lockutils [req-8a4d52b6-8071-4ca2-8ecd-72defceff925 req-b9638ce2-3ba9-4e61-8b7b-b2655fcf5058 service nova] Acquiring lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.345486] env[65121]: DEBUG oslo_concurrency.lockutils [req-8a4d52b6-8071-4ca2-8ecd-72defceff925 req-b9638ce2-3ba9-4e61-8b7b-b2655fcf5058 service nova] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 989.345710] env[65121]: DEBUG oslo_concurrency.lockutils [req-8a4d52b6-8071-4ca2-8ecd-72defceff925 req-b9638ce2-3ba9-4e61-8b7b-b2655fcf5058 service nova] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 989.345873] env[65121]: DEBUG nova.compute.manager [req-8a4d52b6-8071-4ca2-8ecd-72defceff925 req-b9638ce2-3ba9-4e61-8b7b-b2655fcf5058 service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] No waiting events found dispatching network-vif-plugged-24f646d7-31dd-416b-a619-d45d5e0fd1b2 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 989.346165] env[65121]: WARNING nova.compute.manager [req-8a4d52b6-8071-4ca2-8ecd-72defceff925 req-b9638ce2-3ba9-4e61-8b7b-b2655fcf5058 service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Received unexpected event network-vif-plugged-24f646d7-31dd-416b-a619-d45d5e0fd1b2 for instance with vm_state building and task_state spawning. [ 989.387360] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ea0d47-49cd-49ac-983c-2860f80ca7fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.396910] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090003d2-2939-4687-80ed-380fe1dad97a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.431142] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be54c6b-a508-4043-8fb2-7232fc7420ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.439531] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d59a7a4-372a-44a8-809f-dde68dce292f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.453494] env[65121]: DEBUG nova.compute.provider_tree [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.456310] env[65121]: DEBUG nova.network.neutron [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Successfully updated port: 24f646d7-31dd-416b-a619-d45d5e0fd1b2 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 989.957751] env[65121]: DEBUG nova.scheduler.client.report [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 989.961849] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "refresh_cache-e46c88ed-2c2e-432d-987d-0e4c99c28e42" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.962108] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "refresh_cache-e46c88ed-2c2e-432d-987d-0e4c99c28e42" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 989.962298] env[65121]: DEBUG nova.network.neutron [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 990.464712] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.526s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.468143] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.468620] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.476843] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.041s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.478433] env[65121]: INFO nova.compute.claims [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.518800] env[65121]: DEBUG nova.network.neutron [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 990.540290] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.540698] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.609930] env[65121]: WARNING neutronclient.v2_0.client [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.610629] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.610966] env[65121]: WARNING openstack [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.698319] env[65121]: DEBUG nova.network.neutron [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Updating instance_info_cache with network_info: [{"id": "24f646d7-31dd-416b-a619-d45d5e0fd1b2", "address": "fa:16:3e:96:bf:99", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24f646d7-31", "ovs_interfaceid": "24f646d7-31dd-416b-a619-d45d5e0fd1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 990.984562] env[65121]: DEBUG oslo_concurrency.lockutils [None req-560c8137-2c8e-4e43-b5ee-75f0749bf208 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 57.464s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.989211] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 32.955s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.989211] env[65121]: INFO nova.compute.manager [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Unshelving [ 991.200692] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "refresh_cache-e46c88ed-2c2e-432d-987d-0e4c99c28e42" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.201076] env[65121]: DEBUG nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Instance network_info: |[{"id": "24f646d7-31dd-416b-a619-d45d5e0fd1b2", "address": "fa:16:3e:96:bf:99", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24f646d7-31", "ovs_interfaceid": "24f646d7-31dd-416b-a619-d45d5e0fd1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 991.201564] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:bf:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24f646d7-31dd-416b-a619-d45d5e0fd1b2', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.209302] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 991.209560] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.209829] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a1c0edc-af3a-44e9-bd40-5f9ffd28e021 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.232089] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.232089] env[65121]: value = "task-5106931" [ 991.232089] env[65121]: _type = "Task" [ 991.232089] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.241104] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106931, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.377457] env[65121]: DEBUG nova.compute.manager [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Received event network-changed-24f646d7-31dd-416b-a619-d45d5e0fd1b2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 991.377685] env[65121]: DEBUG nova.compute.manager [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Refreshing instance network info cache due to event network-changed-24f646d7-31dd-416b-a619-d45d5e0fd1b2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 991.377922] env[65121]: DEBUG oslo_concurrency.lockutils [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Acquiring lock "refresh_cache-e46c88ed-2c2e-432d-987d-0e4c99c28e42" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.378228] env[65121]: DEBUG oslo_concurrency.lockutils [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Acquired lock "refresh_cache-e46c88ed-2c2e-432d-987d-0e4c99c28e42" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.378319] env[65121]: DEBUG nova.network.neutron [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Refreshing network info cache for port 24f646d7-31dd-416b-a619-d45d5e0fd1b2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 991.741888] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106931, 'name': CreateVM_Task, 'duration_secs': 0.342951} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.744533] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.745619] env[65121]: WARNING neutronclient.v2_0.client [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.745803] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.746061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.746409] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 991.746723] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3205807c-5a57-4662-b731-7641f0c47b6d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.751760] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 991.751760] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cbba40-bfec-99bd-5e71-6943df2759fd" [ 991.751760] env[65121]: _type = "Task" [ 991.751760] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.762657] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cbba40-bfec-99bd-5e71-6943df2759fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.881125] env[65121]: WARNING neutronclient.v2_0.client [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.881795] env[65121]: WARNING openstack [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.882148] env[65121]: WARNING openstack [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.936414] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367bf36e-0a76-420f-9552-701134b28bf4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.948684] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5380a806-98cc-44d2-83cf-f11439fe55a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.004773] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6261be14-1d36-4ba6-9b5c-98820a09d754 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.010229] env[65121]: DEBUG nova.compute.utils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 992.020643] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab6b3d6-48c6-4aaa-aaa2-15207211d09a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.043387] env[65121]: WARNING openstack [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 992.043988] env[65121]: WARNING openstack [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 992.054700] env[65121]: DEBUG nova.compute.provider_tree [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.111284] env[65121]: WARNING neutronclient.v2_0.client [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 992.111969] env[65121]: WARNING openstack [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 992.112320] env[65121]: WARNING openstack [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 992.200285] env[65121]: DEBUG nova.network.neutron [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Updated VIF entry in instance network info cache for port 24f646d7-31dd-416b-a619-d45d5e0fd1b2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 992.200715] env[65121]: DEBUG nova.network.neutron [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Updating instance_info_cache with network_info: [{"id": "24f646d7-31dd-416b-a619-d45d5e0fd1b2", "address": "fa:16:3e:96:bf:99", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24f646d7-31", "ovs_interfaceid": "24f646d7-31dd-416b-a619-d45d5e0fd1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 992.263721] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cbba40-bfec-99bd-5e71-6943df2759fd, 'name': SearchDatastore_Task, 'duration_secs': 0.010256} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.263925] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.264138] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.264372] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.264511] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 992.264691] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.264952] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d89d0a0-8dd2-45ea-8753-e0a436fe2d10 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.273599] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.273811] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 992.274517] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24f896e8-1e7a-4082-a608-0c5b29c2c6c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.280296] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 992.280296] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527c89e0-5ae1-a19f-0971-c028b6504915" [ 992.280296] env[65121]: _type = "Task" [ 992.280296] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.288479] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527c89e0-5ae1-a19f-0971-c028b6504915, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.515104] env[65121]: INFO nova.virt.block_device [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Booting with volume 02de1e93-4004-47b3-a1b4-d2690ef0da94 at /dev/sdb [ 992.552794] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a9c5583-ce6a-4b33-8bf1-e1d413d32995 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.558290] env[65121]: DEBUG nova.scheduler.client.report [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 992.566152] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbdb0d66-8fca-4d9f-b9f7-96c1276e2845 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.600318] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6876c7d3-ecf3-4349-a904-92680c50324a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.610158] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a0127f-6578-4e40-984a-4aca2f931fe9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.645952] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b80ddc0-0bd0-4b75-869f-64b6a50028fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.653486] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb4f228-87a5-4d70-a673-73bc8ed40ec0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.669427] env[65121]: DEBUG nova.virt.block_device [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating existing volume attachment record: bd51ab0a-aead-4779-9adc-0db1b6b8572c {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 992.703674] env[65121]: DEBUG oslo_concurrency.lockutils [req-6e252788-ab65-47b1-a657-4c3a0ae57553 req-951a7f53-74e2-486d-b064-a24a5d5b178f service nova] Releasing lock "refresh_cache-e46c88ed-2c2e-432d-987d-0e4c99c28e42" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.791307] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527c89e0-5ae1-a19f-0971-c028b6504915, 'name': SearchDatastore_Task, 'duration_secs': 0.009127} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.792141] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fcc6f7e-e347-4f85-aed3-6e2236fffc9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.798270] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 992.798270] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52089676-dafb-1fd9-5c8d-8ccc6313b7d8" [ 992.798270] env[65121]: _type = "Task" [ 992.798270] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.806664] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52089676-dafb-1fd9-5c8d-8ccc6313b7d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.065039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.065664] env[65121]: DEBUG nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 993.068629] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.452s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.070278] env[65121]: INFO nova.compute.claims [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.310277] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52089676-dafb-1fd9-5c8d-8ccc6313b7d8, 'name': SearchDatastore_Task, 'duration_secs': 0.023814} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.310450] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 993.310677] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e46c88ed-2c2e-432d-987d-0e4c99c28e42/e46c88ed-2c2e-432d-987d-0e4c99c28e42.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 993.311358] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69becfbd-1686-4a67-bf28-8b4f3fc33338 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.318574] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 993.318574] env[65121]: value = "task-5106935" [ 993.318574] env[65121]: _type = "Task" [ 993.318574] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.327584] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106935, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.577363] env[65121]: DEBUG nova.compute.utils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 993.579116] env[65121]: DEBUG nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 993.579397] env[65121]: DEBUG nova.network.neutron [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 993.579858] env[65121]: WARNING neutronclient.v2_0.client [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.580318] env[65121]: WARNING neutronclient.v2_0.client [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.581030] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.581457] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.649400] env[65121]: DEBUG nova.policy [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '578dbf589da24c04968e104a4db5accd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c0ac05495754323834faaad28ff13d8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 993.829724] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106935, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478667} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.830124] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e46c88ed-2c2e-432d-987d-0e4c99c28e42/e46c88ed-2c2e-432d-987d-0e4c99c28e42.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.830124] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.830389] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ca73287-b6e9-4d0f-853e-a45c1ee3990e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.839042] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 993.839042] env[65121]: value = "task-5106936" [ 993.839042] env[65121]: _type = "Task" [ 993.839042] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.850404] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106936, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.952330] env[65121]: DEBUG nova.network.neutron [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Successfully created port: d9f74553-cfad-41e7-adf0-5faf0b793cf0 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 994.091330] env[65121]: DEBUG nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 994.350435] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106936, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068108} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.353403] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.354462] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78f4780-fbda-4092-bd10-b990b5d57f77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.378345] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] e46c88ed-2c2e-432d-987d-0e4c99c28e42/e46c88ed-2c2e-432d-987d-0e4c99c28e42.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.381374] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51a73416-afcd-49c2-9517-462815f36eb5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.405543] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 994.405543] env[65121]: value = "task-5106937" [ 994.405543] env[65121]: _type = "Task" [ 994.405543] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.417559] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106937, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.543172] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b5e33e-ca89-40d7-839b-ee85236cf261 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.551556] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2dbbe7d-c527-42bc-892d-1821df2d9374 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.584077] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c77d6a-90ea-4ba9-aee0-09b8775c1c65 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.592640] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef487a93-210d-4ff5-bbf4-792e65446498 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.611452] env[65121]: DEBUG nova.compute.provider_tree [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.916076] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106937, 'name': ReconfigVM_Task, 'duration_secs': 0.45968} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.916370] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Reconfigured VM instance instance-00000052 to attach disk [datastore1] e46c88ed-2c2e-432d-987d-0e4c99c28e42/e46c88ed-2c2e-432d-987d-0e4c99c28e42.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.917188] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fbd06e5e-d251-4281-96bf-10884bb24593 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.924183] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 994.924183] env[65121]: value = "task-5106938" [ 994.924183] env[65121]: _type = "Task" [ 994.924183] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.934592] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106938, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.103321] env[65121]: DEBUG nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 995.114918] env[65121]: DEBUG nova.scheduler.client.report [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.131744] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 995.131985] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 995.132152] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 995.132332] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 995.132470] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 995.132609] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 995.132808] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.132961] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 995.133137] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 995.133358] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 995.133988] env[65121]: DEBUG nova.virt.hardware [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 995.134544] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad8f7a8-a447-42ca-9f9f-22afcd157ffa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.145165] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7e931f-7a72-4d37-b8b4-ac22262fdb12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.433894] env[65121]: DEBUG nova.compute.manager [req-db120282-e535-489c-97f8-5a1060a9bae0 req-c8b1db57-fdbd-4122-9c7e-c1aae620e8ae service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Received event network-vif-plugged-d9f74553-cfad-41e7-adf0-5faf0b793cf0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 995.434009] env[65121]: DEBUG oslo_concurrency.lockutils [req-db120282-e535-489c-97f8-5a1060a9bae0 req-c8b1db57-fdbd-4122-9c7e-c1aae620e8ae service nova] Acquiring lock "17f34d32-e690-40d8-99a4-acf53fa859c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.434220] env[65121]: DEBUG oslo_concurrency.lockutils [req-db120282-e535-489c-97f8-5a1060a9bae0 req-c8b1db57-fdbd-4122-9c7e-c1aae620e8ae service nova] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.434386] env[65121]: DEBUG oslo_concurrency.lockutils [req-db120282-e535-489c-97f8-5a1060a9bae0 req-c8b1db57-fdbd-4122-9c7e-c1aae620e8ae service nova] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.434546] env[65121]: DEBUG nova.compute.manager [req-db120282-e535-489c-97f8-5a1060a9bae0 req-c8b1db57-fdbd-4122-9c7e-c1aae620e8ae service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] No waiting events found dispatching network-vif-plugged-d9f74553-cfad-41e7-adf0-5faf0b793cf0 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 995.434702] env[65121]: WARNING nova.compute.manager [req-db120282-e535-489c-97f8-5a1060a9bae0 req-c8b1db57-fdbd-4122-9c7e-c1aae620e8ae service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Received unexpected event network-vif-plugged-d9f74553-cfad-41e7-adf0-5faf0b793cf0 for instance with vm_state building and task_state spawning. [ 995.438574] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106938, 'name': Rename_Task, 'duration_secs': 0.191581} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.438853] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 995.439116] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08b50749-e86d-484c-9e2e-56afa3eaa27e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.447351] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 995.447351] env[65121]: value = "task-5106940" [ 995.447351] env[65121]: _type = "Task" [ 995.447351] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.457144] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.540544] env[65121]: DEBUG nova.network.neutron [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Successfully updated port: d9f74553-cfad-41e7-adf0-5faf0b793cf0 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 995.621658] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.622430] env[65121]: DEBUG nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 995.626328] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.270s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.626676] env[65121]: DEBUG nova.objects.instance [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lazy-loading 'resources' on Instance uuid 59d6005d-6caf-4898-b791-70d9c015cdb4 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.958930] env[65121]: DEBUG oslo_vmware.api [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106940, 'name': PowerOnVM_Task, 'duration_secs': 0.488682} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.959470] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.959470] env[65121]: INFO nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Took 6.99 seconds to spawn the instance on the hypervisor. [ 995.959635] env[65121]: DEBUG nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 995.960524] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88adfafe-d089-40c8-bf44-962c3b849836 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.046985] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "refresh_cache-17f34d32-e690-40d8-99a4-acf53fa859c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.047351] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquired lock "refresh_cache-17f34d32-e690-40d8-99a4-acf53fa859c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.047351] env[65121]: DEBUG nova.network.neutron [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 996.130511] env[65121]: DEBUG nova.compute.utils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 996.132250] env[65121]: DEBUG nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 996.132444] env[65121]: DEBUG nova.network.neutron [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 996.132909] env[65121]: WARNING neutronclient.v2_0.client [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.133090] env[65121]: WARNING neutronclient.v2_0.client [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.134035] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.134381] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.146417] env[65121]: DEBUG nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 996.200776] env[65121]: DEBUG nova.policy [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '175d752afbc745258500be8624bfbf14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e491b88dcaad42dcb345b3217bc6f941', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 996.488108] env[65121]: INFO nova.compute.manager [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Took 44.89 seconds to build instance. [ 996.538266] env[65121]: DEBUG nova.network.neutron [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Successfully created port: db0cba48-a3d6-4aff-b44f-c1e6672a1daf {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 996.543619] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c16fbd-952b-46a3-b1b3-ef91e058c388 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.551779] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.552244] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.560283] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c162f18-3aab-4f16-b442-cb3089125778 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.593033] env[65121]: DEBUG nova.network.neutron [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 996.596104] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7154357d-6c53-4660-b79f-e9307e7237e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.604065] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ff493f-ebc7-44d8-a912-1a4198c6f32c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.618119] env[65121]: DEBUG nova.compute.provider_tree [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.673705] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.674171] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.773683] env[65121]: WARNING neutronclient.v2_0.client [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.774412] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.774744] env[65121]: WARNING openstack [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.891725] env[65121]: DEBUG nova.network.neutron [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Updating instance_info_cache with network_info: [{"id": "d9f74553-cfad-41e7-adf0-5faf0b793cf0", "address": "fa:16:3e:fa:ed:5d", "network": {"id": "f50d6a93-15b7-46b8-ae6f-15011303731c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2081068655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c0ac05495754323834faaad28ff13d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f74553-cf", "ovs_interfaceid": "d9f74553-cfad-41e7-adf0-5faf0b793cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 996.991725] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7ea81f78-a7ea-46f4-bcfe-cd8455942982 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.403s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.120976] env[65121]: DEBUG nova.scheduler.client.report [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 997.160732] env[65121]: DEBUG nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 997.187677] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 997.187912] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 997.188079] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 997.188294] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 997.188447] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 997.188598] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 997.188856] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.189034] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 997.189203] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 997.189384] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 997.189655] env[65121]: DEBUG nova.virt.hardware [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 997.190568] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b38765-48c5-49de-9a04-b90dcc52ffc4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.200028] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd371e8-5866-40af-8cda-480dec210a03 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.320405] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.320765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.320982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.321183] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.321371] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.323636] env[65121]: INFO nova.compute.manager [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Terminating instance [ 997.394271] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Releasing lock "refresh_cache-17f34d32-e690-40d8-99a4-acf53fa859c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.394566] env[65121]: DEBUG nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Instance network_info: |[{"id": "d9f74553-cfad-41e7-adf0-5faf0b793cf0", "address": "fa:16:3e:fa:ed:5d", "network": {"id": "f50d6a93-15b7-46b8-ae6f-15011303731c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2081068655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c0ac05495754323834faaad28ff13d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f74553-cf", "ovs_interfaceid": "d9f74553-cfad-41e7-adf0-5faf0b793cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 997.395036] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:ed:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20641d67-1612-4b9c-8924-7a77df9c8e6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9f74553-cfad-41e7-adf0-5faf0b793cf0', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.402837] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Creating folder: Project (1c0ac05495754323834faaad28ff13d8). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 997.403473] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ebc942a-3795-4dfa-8aa0-0cd7c2d48d92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.414987] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Created folder: Project (1c0ac05495754323834faaad28ff13d8) in parent group-v993268. [ 997.415205] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Creating folder: Instances. Parent ref: group-v993501. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 997.415471] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ba0872c-01c9-4d07-89c9-425c0752f12b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.424788] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Created folder: Instances in parent group-v993501. [ 997.425013] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 997.425211] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 997.425433] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bc8ef83-ff7f-4170-b5da-9d048dd1602d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.445793] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.445793] env[65121]: value = "task-5106943" [ 997.445793] env[65121]: _type = "Task" [ 997.445793] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.455324] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106943, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.466205] env[65121]: DEBUG nova.compute.manager [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Received event network-changed-d9f74553-cfad-41e7-adf0-5faf0b793cf0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 997.466467] env[65121]: DEBUG nova.compute.manager [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Refreshing instance network info cache due to event network-changed-d9f74553-cfad-41e7-adf0-5faf0b793cf0. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 997.466691] env[65121]: DEBUG oslo_concurrency.lockutils [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Acquiring lock "refresh_cache-17f34d32-e690-40d8-99a4-acf53fa859c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.466825] env[65121]: DEBUG oslo_concurrency.lockutils [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Acquired lock "refresh_cache-17f34d32-e690-40d8-99a4-acf53fa859c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.466981] env[65121]: DEBUG nova.network.neutron [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Refreshing network info cache for port d9f74553-cfad-41e7-adf0-5faf0b793cf0 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 997.492705] env[65121]: DEBUG nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 997.626893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.629615] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.039s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.629914] env[65121]: DEBUG nova.objects.instance [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lazy-loading 'resources' on Instance uuid 68e3e4e6-6ca1-4f76-a4af-5112a2042c61 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.653673] env[65121]: INFO nova.scheduler.client.report [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted allocations for instance 59d6005d-6caf-4898-b791-70d9c015cdb4 [ 997.827143] env[65121]: DEBUG nova.compute.manager [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 997.827416] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.828772] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054052a8-8b4b-4bcb-9003-7bb035d064a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.836310] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.836540] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ee7d38d-8ec3-4fad-a0dd-8599e63f5dde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.842380] env[65121]: DEBUG oslo_vmware.api [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 997.842380] env[65121]: value = "task-5106944" [ 997.842380] env[65121]: _type = "Task" [ 997.842380] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.850809] env[65121]: DEBUG oslo_vmware.api [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.956972] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106943, 'name': CreateVM_Task, 'duration_secs': 0.345372} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.957160] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 997.957710] env[65121]: WARNING neutronclient.v2_0.client [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 997.958097] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.958256] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.958589] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 997.958858] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9575d525-c8c6-4007-aebc-9c4db74e76ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.964632] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 997.964632] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e44256-9278-2c91-b1ef-172ba48aed14" [ 997.964632] env[65121]: _type = "Task" [ 997.964632] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.970316] env[65121]: WARNING neutronclient.v2_0.client [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 997.971055] env[65121]: WARNING openstack [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.971441] env[65121]: WARNING openstack [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.985759] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e44256-9278-2c91-b1ef-172ba48aed14, 'name': SearchDatastore_Task, 'duration_secs': 0.011081} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.986438] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.986438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.986783] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.987051] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.987357] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.987911] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be216061-afb7-4ebc-b648-229658ba1ca8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.008185] env[65121]: DEBUG nova.compute.manager [req-a8e1f48c-8dfc-447c-b4d3-ed5fd7b74fab req-6a9a1d81-0fc4-449f-8031-47315b56e090 service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Received event network-vif-plugged-db0cba48-a3d6-4aff-b44f-c1e6672a1daf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 998.008608] env[65121]: DEBUG oslo_concurrency.lockutils [req-a8e1f48c-8dfc-447c-b4d3-ed5fd7b74fab req-6a9a1d81-0fc4-449f-8031-47315b56e090 service nova] Acquiring lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.009538] env[65121]: DEBUG oslo_concurrency.lockutils [req-a8e1f48c-8dfc-447c-b4d3-ed5fd7b74fab req-6a9a1d81-0fc4-449f-8031-47315b56e090 service nova] Lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.009538] env[65121]: DEBUG oslo_concurrency.lockutils [req-a8e1f48c-8dfc-447c-b4d3-ed5fd7b74fab req-6a9a1d81-0fc4-449f-8031-47315b56e090 service nova] Lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.009538] env[65121]: DEBUG nova.compute.manager [req-a8e1f48c-8dfc-447c-b4d3-ed5fd7b74fab req-6a9a1d81-0fc4-449f-8031-47315b56e090 service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] No waiting events found dispatching network-vif-plugged-db0cba48-a3d6-4aff-b44f-c1e6672a1daf {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 998.009538] env[65121]: WARNING nova.compute.manager [req-a8e1f48c-8dfc-447c-b4d3-ed5fd7b74fab req-6a9a1d81-0fc4-449f-8031-47315b56e090 service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Received unexpected event network-vif-plugged-db0cba48-a3d6-4aff-b44f-c1e6672a1daf for instance with vm_state building and task_state spawning. [ 998.011351] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.011351] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 998.012677] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d366012-117c-460c-9156-40457c5eccb1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.021485] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.022049] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 998.022049] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5203853b-e83d-a929-0f2c-ad5a27531fae" [ 998.022049] env[65121]: _type = "Task" [ 998.022049] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.032561] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5203853b-e83d-a929-0f2c-ad5a27531fae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.074978] env[65121]: DEBUG nova.network.neutron [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Successfully updated port: db0cba48-a3d6-4aff-b44f-c1e6672a1daf {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 998.094550] env[65121]: WARNING openstack [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.094927] env[65121]: WARNING openstack [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.161526] env[65121]: WARNING neutronclient.v2_0.client [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.162164] env[65121]: WARNING openstack [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.162529] env[65121]: WARNING openstack [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.172732] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1732a3ff-b5f4-4dc2-b1aa-79d007e931ba tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "59d6005d-6caf-4898-b791-70d9c015cdb4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.851s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.272387] env[65121]: DEBUG nova.network.neutron [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Updated VIF entry in instance network info cache for port d9f74553-cfad-41e7-adf0-5faf0b793cf0. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 998.272767] env[65121]: DEBUG nova.network.neutron [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Updating instance_info_cache with network_info: [{"id": "d9f74553-cfad-41e7-adf0-5faf0b793cf0", "address": "fa:16:3e:fa:ed:5d", "network": {"id": "f50d6a93-15b7-46b8-ae6f-15011303731c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-2081068655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c0ac05495754323834faaad28ff13d8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9f74553-cf", "ovs_interfaceid": "d9f74553-cfad-41e7-adf0-5faf0b793cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 998.353210] env[65121]: DEBUG oslo_vmware.api [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106944, 'name': PowerOffVM_Task, 'duration_secs': 0.19244} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.353503] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.353663] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.354033] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e990401-f17b-4727-aa96-16d3dab5e5a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.426509] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.426509] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.426650] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleting the datastore file [datastore1] e46c88ed-2c2e-432d-987d-0e4c99c28e42 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.431327] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9bf3eca6-ef10-4c05-b3e5-36afa74095e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.440070] env[65121]: DEBUG oslo_vmware.api [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 998.440070] env[65121]: value = "task-5106946" [ 998.440070] env[65121]: _type = "Task" [ 998.440070] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.451882] env[65121]: DEBUG oslo_vmware.api [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.532168] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5203853b-e83d-a929-0f2c-ad5a27531fae, 'name': SearchDatastore_Task, 'duration_secs': 0.022758} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.534330] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a2ddd4-ece6-4491-abb5-a6bfef24c9cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.536840] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f20c146-5648-472d-b8e2-1a8c64737bfe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.545590] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8585dc75-9a5d-4c43-a438-aa1d15ff5217 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.548876] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 998.548876] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5273ce98-38d1-ac72-ace6-b50dd9021e4f" [ 998.548876] env[65121]: _type = "Task" [ 998.548876] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.579457] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.579662] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 998.579818] env[65121]: DEBUG nova.network.neutron [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 998.583928] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926fc98d-3276-4a78-8fac-58a28cbfeb1c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.590467] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5273ce98-38d1-ac72-ace6-b50dd9021e4f, 'name': SearchDatastore_Task, 'duration_secs': 0.012664} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.591842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.592177] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 17f34d32-e690-40d8-99a4-acf53fa859c6/17f34d32-e690-40d8-99a4-acf53fa859c6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 998.592727] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b89feeb-1a8b-4523-a8f1-27f5bc60f94d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.599177] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf74da06-b56c-456c-8726-132ab16696ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.604650] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 998.604650] env[65121]: value = "task-5106947" [ 998.604650] env[65121]: _type = "Task" [ 998.604650] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.616959] env[65121]: DEBUG nova.compute.provider_tree [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.624425] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.776757] env[65121]: DEBUG oslo_concurrency.lockutils [req-0d9d34fe-c58a-40c1-8c97-92b7fffaf23f req-9a4e6ded-1ff3-418c-afbf-ae96b29e0705 service nova] Releasing lock "refresh_cache-17f34d32-e690-40d8-99a4-acf53fa859c6" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.786482] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.950802] env[65121]: DEBUG oslo_vmware.api [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5106946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229046} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.951087] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.951327] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 998.951535] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.951734] env[65121]: INFO nova.compute.manager [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Took 1.12 seconds to destroy the instance on the hypervisor. [ 998.952020] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 998.952369] env[65121]: DEBUG nova.compute.manager [-] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 998.952369] env[65121]: DEBUG nova.network.neutron [-] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 998.952707] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.953289] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.953551] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.994112] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.083130] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.083695] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.118988] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106947, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514993} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.120058] env[65121]: DEBUG nova.scheduler.client.report [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 999.123734] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 17f34d32-e690-40d8-99a4-acf53fa859c6/17f34d32-e690-40d8-99a4-acf53fa859c6.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.127024] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.127024] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85c9adf6-ea2a-43c5-90b2-b71e5b9da483 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.134393] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 999.134393] env[65121]: value = "task-5106948" [ 999.134393] env[65121]: _type = "Task" [ 999.134393] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.148658] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106948, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.149897] env[65121]: DEBUG nova.network.neutron [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 999.257766] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.258214] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.343591] env[65121]: WARNING neutronclient.v2_0.client [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.344303] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.344659] env[65121]: WARNING openstack [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.442621] env[65121]: DEBUG nova.network.neutron [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance_info_cache with network_info: [{"id": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "address": "fa:16:3e:d9:2e:bc", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb0cba48-a3", "ovs_interfaceid": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 999.627020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.997s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.629802] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.943s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.631502] env[65121]: INFO nova.compute.claims [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.644725] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106948, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078204} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.645405] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.646028] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474f0797-0cbb-4da4-997d-ff03798055c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.673340] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 17f34d32-e690-40d8-99a4-acf53fa859c6/17f34d32-e690-40d8-99a4-acf53fa859c6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.674671] env[65121]: INFO nova.scheduler.client.report [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Deleted allocations for instance 68e3e4e6-6ca1-4f76-a4af-5112a2042c61 [ 999.675937] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da054b8e-42a6-4985-ab91-f2d27c60f856 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.704038] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 999.704038] env[65121]: value = "task-5106949" [ 999.704038] env[65121]: _type = "Task" [ 999.704038] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.713295] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.737109] env[65121]: DEBUG nova.network.neutron [-] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 999.946786] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 999.947246] env[65121]: DEBUG nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Instance network_info: |[{"id": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "address": "fa:16:3e:d9:2e:bc", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb0cba48-a3", "ovs_interfaceid": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 999.947858] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:2e:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'db0cba48-a3d6-4aff-b44f-c1e6672a1daf', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.955624] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 999.955891] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 999.956138] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4c50830-ffb1-4734-ad0c-72d74fae817a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.976993] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.976993] env[65121]: value = "task-5106950" [ 999.976993] env[65121]: _type = "Task" [ 999.976993] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.984903] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106950, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.043220] env[65121]: DEBUG nova.compute.manager [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Received event network-changed-db0cba48-a3d6-4aff-b44f-c1e6672a1daf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1000.043485] env[65121]: DEBUG nova.compute.manager [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Refreshing instance network info cache due to event network-changed-db0cba48-a3d6-4aff-b44f-c1e6672a1daf. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1000.043704] env[65121]: DEBUG oslo_concurrency.lockutils [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Acquiring lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.043878] env[65121]: DEBUG oslo_concurrency.lockutils [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Acquired lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1000.044102] env[65121]: DEBUG nova.network.neutron [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Refreshing network info cache for port db0cba48-a3d6-4aff-b44f-c1e6672a1daf {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1000.203017] env[65121]: DEBUG oslo_concurrency.lockutils [None req-640628f0-7bba-4796-add3-d02017275739 tempest-ServerShowV247Test-2063705043 tempest-ServerShowV247Test-2063705043-project-member] Lock "68e3e4e6-6ca1-4f76-a4af-5112a2042c61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.991s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.214477] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106949, 'name': ReconfigVM_Task, 'duration_secs': 0.299557} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.214805] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 17f34d32-e690-40d8-99a4-acf53fa859c6/17f34d32-e690-40d8-99a4-acf53fa859c6.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.215569] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39c29f0e-1881-4613-af78-9705a8b8b7e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.224742] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 1000.224742] env[65121]: value = "task-5106951" [ 1000.224742] env[65121]: _type = "Task" [ 1000.224742] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.236833] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106951, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.239510] env[65121]: INFO nova.compute.manager [-] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Took 1.29 seconds to deallocate network for instance. [ 1000.489023] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106950, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.546908] env[65121]: WARNING neutronclient.v2_0.client [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1000.547743] env[65121]: WARNING openstack [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1000.548229] env[65121]: WARNING openstack [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1000.652307] env[65121]: WARNING openstack [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1000.652831] env[65121]: WARNING openstack [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1000.733586] env[65121]: WARNING neutronclient.v2_0.client [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1000.734282] env[65121]: WARNING openstack [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1000.734644] env[65121]: WARNING openstack [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1000.749987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.750334] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106951, 'name': Rename_Task, 'duration_secs': 0.163209} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.750647] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.750990] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f3a8259-8cd2-4f1c-bcf7-3547071e5f63 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.758314] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 1000.758314] env[65121]: value = "task-5106952" [ 1000.758314] env[65121]: _type = "Task" [ 1000.758314] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.768645] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.828440] env[65121]: DEBUG nova.network.neutron [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updated VIF entry in instance network info cache for port db0cba48-a3d6-4aff-b44f-c1e6672a1daf. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1000.828825] env[65121]: DEBUG nova.network.neutron [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance_info_cache with network_info: [{"id": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "address": "fa:16:3e:d9:2e:bc", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb0cba48-a3", "ovs_interfaceid": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1000.992347] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106950, 'name': CreateVM_Task, 'duration_secs': 0.534057} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.992698] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1000.994029] env[65121]: WARNING neutronclient.v2_0.client [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1000.994029] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.994147] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1000.994408] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1000.997291] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e83e1584-6905-4166-a12e-a9b40345b06d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.004306] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1001.004306] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521ed1cb-d66d-21ce-febf-2dc32538443f" [ 1001.004306] env[65121]: _type = "Task" [ 1001.004306] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.013505] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521ed1cb-d66d-21ce-febf-2dc32538443f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.030749] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813400e6-c204-4b99-9838-d37ffbca764e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.041138] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13efdc7-28f9-4d26-99af-000988527006 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.075563] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cf4a47-1cbd-4237-ac82-ab600a5cbc78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.086036] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f288641-2a22-43e3-a3b1-a431858bbe22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.100831] env[65121]: DEBUG nova.compute.provider_tree [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.269884] env[65121]: DEBUG oslo_vmware.api [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106952, 'name': PowerOnVM_Task, 'duration_secs': 0.496439} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.270271] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.270508] env[65121]: INFO nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Took 6.17 seconds to spawn the instance on the hypervisor. [ 1001.270731] env[65121]: DEBUG nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1001.271693] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a38ccf-5601-46cf-bef2-b11b067f44a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.331443] env[65121]: DEBUG oslo_concurrency.lockutils [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] Releasing lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1001.331721] env[65121]: DEBUG nova.compute.manager [req-718aa088-be87-41d6-a2f4-3286d7675b6f req-57fcaa8d-362d-4e3e-b68d-8de4ff5df59c service nova] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Received event network-vif-deleted-24f646d7-31dd-416b-a619-d45d5e0fd1b2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1001.514937] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521ed1cb-d66d-21ce-febf-2dc32538443f, 'name': SearchDatastore_Task, 'duration_secs': 0.010952} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.515299] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1001.515612] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.515968] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.516281] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1001.517102] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.517102] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e9e6f65-3a44-4ded-9d9c-ab7ff93410ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.527029] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.527222] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1001.527903] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f56b66c2-ea5e-450d-b9a3-e9adf89c96b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.533239] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1001.533239] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5274651d-ece2-a9c9-394f-30ea4241d485" [ 1001.533239] env[65121]: _type = "Task" [ 1001.533239] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.541599] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5274651d-ece2-a9c9-394f-30ea4241d485, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.603935] env[65121]: DEBUG nova.scheduler.client.report [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1001.791134] env[65121]: INFO nova.compute.manager [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Took 46.38 seconds to build instance. [ 1001.982440] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "17f34d32-e690-40d8-99a4-acf53fa859c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.051186] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5274651d-ece2-a9c9-394f-30ea4241d485, 'name': SearchDatastore_Task, 'duration_secs': 0.00973} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.052044] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1267495-2f46-4258-9c11-41ef73f7be29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.059211] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1002.059211] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520e92bd-14a6-0aec-32fb-0a8f8e1dbb49" [ 1002.059211] env[65121]: _type = "Task" [ 1002.059211] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.067932] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520e92bd-14a6-0aec-32fb-0a8f8e1dbb49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.109254] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.109836] env[65121]: DEBUG nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1002.112573] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.865s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.114175] env[65121]: INFO nova.compute.claims [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.292961] env[65121]: DEBUG oslo_concurrency.lockutils [None req-56bc9b41-8a3f-438c-9bb6-15b32110b1d3 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.895s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.294405] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.312s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.294621] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "17f34d32-e690-40d8-99a4-acf53fa859c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.294837] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.295006] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.296947] env[65121]: INFO nova.compute.manager [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Terminating instance [ 1002.570516] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520e92bd-14a6-0aec-32fb-0a8f8e1dbb49, 'name': SearchDatastore_Task, 'duration_secs': 0.010027} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.570771] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1002.571044] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 564159fe-6ecd-4276-8d65-cbe25859493a/564159fe-6ecd-4276-8d65-cbe25859493a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1002.571305] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6c312a0-2bf3-4c0e-84f6-24d1341ebca4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.577878] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1002.577878] env[65121]: value = "task-5106953" [ 1002.577878] env[65121]: _type = "Task" [ 1002.577878] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.586017] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106953, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.619034] env[65121]: DEBUG nova.compute.utils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1002.622607] env[65121]: DEBUG nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1002.622828] env[65121]: DEBUG nova.network.neutron [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1002.623208] env[65121]: WARNING neutronclient.v2_0.client [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.623544] env[65121]: WARNING neutronclient.v2_0.client [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.624184] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.624550] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.708155] env[65121]: DEBUG nova.policy [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1c51fc43b7f44ce9b023d8cb769e029', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89110618b6ef4ce68b2ca53e7e168139', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1002.797049] env[65121]: DEBUG nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1002.800289] env[65121]: DEBUG nova.compute.manager [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1002.800491] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.801413] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0528839d-a903-422a-b488-93a82cd85681 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.813821] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.816823] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df4910a3-4ad7-49c1-83cc-0de4d282c956 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.827268] env[65121]: DEBUG oslo_vmware.api [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 1002.827268] env[65121]: value = "task-5106954" [ 1002.827268] env[65121]: _type = "Task" [ 1002.827268] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.839199] env[65121]: DEBUG oslo_vmware.api [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.089324] env[65121]: DEBUG nova.network.neutron [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Successfully created port: adb6945c-5a1e-487c-b116-4200874d80a0 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1003.099106] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106953, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49839} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.099106] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 564159fe-6ecd-4276-8d65-cbe25859493a/564159fe-6ecd-4276-8d65-cbe25859493a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.099106] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.099408] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-505acea5-1ddc-456a-853e-94c6eecc5af3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.104054] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a85417b-0d10-4470-ad70-52b9f16e16aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.113639] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b57ed6-432f-4bb7-816d-ada6d1eb1357 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.117169] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1003.117169] env[65121]: value = "task-5106955" [ 1003.117169] env[65121]: _type = "Task" [ 1003.117169] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.146572] env[65121]: DEBUG nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1003.154443] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72bef68-a328-4758-ad84-cd1e05f05a6c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.160797] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106955, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.167760] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf8e203d-5047-4d97-856e-7a3b2afd81ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.184145] env[65121]: DEBUG nova.compute.provider_tree [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.332148] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.340456] env[65121]: DEBUG oslo_vmware.api [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106954, 'name': PowerOffVM_Task, 'duration_secs': 0.244537} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.340456] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.340456] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.340456] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea457fc4-49ac-4fbf-83f4-8e100e88ed4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.430049] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.430278] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.430437] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Deleting the datastore file [datastore1] 17f34d32-e690-40d8-99a4-acf53fa859c6 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.430678] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c648d1f-a0c2-4592-8097-c88add8f1ba7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.438375] env[65121]: DEBUG oslo_vmware.api [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for the task: (returnval){ [ 1003.438375] env[65121]: value = "task-5106957" [ 1003.438375] env[65121]: _type = "Task" [ 1003.438375] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.447165] env[65121]: DEBUG oslo_vmware.api [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106957, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.513736] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.514274] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.514964] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.515084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.515251] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.517735] env[65121]: INFO nova.compute.manager [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Terminating instance [ 1003.626960] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106955, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065741} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.627510] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.628345] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78f8ed2-f7a1-4c0d-b2bf-7083d71a2d3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.651323] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 564159fe-6ecd-4276-8d65-cbe25859493a/564159fe-6ecd-4276-8d65-cbe25859493a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.651656] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d66067fd-2e61-481c-bd41-53951f42f03e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.677307] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1003.677307] env[65121]: value = "task-5106958" [ 1003.677307] env[65121]: _type = "Task" [ 1003.677307] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.686702] env[65121]: DEBUG nova.scheduler.client.report [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1003.690842] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.949205] env[65121]: DEBUG oslo_vmware.api [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Task: {'id': task-5106957, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168086} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.949463] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.949659] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.949829] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.950037] env[65121]: INFO nova.compute.manager [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1003.950291] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1003.950482] env[65121]: DEBUG nova.compute.manager [-] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1003.950577] env[65121]: DEBUG nova.network.neutron [-] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1003.950822] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.951376] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.951627] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.990766] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1004.021844] env[65121]: DEBUG nova.compute.manager [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1004.021981] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1004.023045] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf035154-e1a6-416a-831a-d66e7d540f35 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.031860] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.032153] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53ce5376-4e42-40cc-996e-d8ec35f4384b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.039935] env[65121]: DEBUG oslo_vmware.api [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 1004.039935] env[65121]: value = "task-5106959" [ 1004.039935] env[65121]: _type = "Task" [ 1004.039935] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.050790] env[65121]: DEBUG oslo_vmware.api [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106959, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.173500] env[65121]: DEBUG nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1004.187022] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106958, 'name': ReconfigVM_Task, 'duration_secs': 0.296282} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.187426] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 564159fe-6ecd-4276-8d65-cbe25859493a/564159fe-6ecd-4276-8d65-cbe25859493a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.188020] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1688b793-4f98-4a4a-811d-13609b9c45ea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.191965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.192479] env[65121]: DEBUG nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1004.198514] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.878s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.198818] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.200994] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.682s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.203944] env[65121]: INFO nova.compute.claims [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1004.205436] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1004.205436] env[65121]: value = "task-5106960" [ 1004.205436] env[65121]: _type = "Task" [ 1004.205436] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.215732] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1004.216408] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1004.216408] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1004.216408] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1004.216566] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1004.216601] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1004.216788] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1004.216979] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1004.217131] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1004.217288] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1004.217450] env[65121]: DEBUG nova.virt.hardware [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1004.218351] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dfb77a-a2ad-4018-b15c-a03aff237251 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.231155] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485e8a92-0951-4704-bcdb-7d03a8258c4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.235822] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106960, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.238115] env[65121]: INFO nova.scheduler.client.report [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted allocations for instance 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c [ 1004.354516] env[65121]: DEBUG nova.compute.manager [req-0e37a9ef-746d-484a-a2a1-8573a3148129 req-c010d596-a037-4ab8-9901-5e61a0f6c008 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Received event network-vif-deleted-d9f74553-cfad-41e7-adf0-5faf0b793cf0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1004.354757] env[65121]: INFO nova.compute.manager [req-0e37a9ef-746d-484a-a2a1-8573a3148129 req-c010d596-a037-4ab8-9901-5e61a0f6c008 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Neutron deleted interface d9f74553-cfad-41e7-adf0-5faf0b793cf0; detaching it from the instance and deleting it from the info cache [ 1004.354793] env[65121]: DEBUG nova.network.neutron [req-0e37a9ef-746d-484a-a2a1-8573a3148129 req-c010d596-a037-4ab8-9901-5e61a0f6c008 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.551853] env[65121]: DEBUG oslo_vmware.api [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106959, 'name': PowerOffVM_Task, 'duration_secs': 0.231848} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.551853] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.551853] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1004.551853] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19ffee3b-410b-4b78-be50-71fb9f026139 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.599097] env[65121]: DEBUG nova.compute.manager [req-bbdfd07f-bcb6-4621-b114-e1c00e6bd709 req-18e29b5e-1819-43bc-a29f-6944766951d6 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Received event network-vif-plugged-adb6945c-5a1e-487c-b116-4200874d80a0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1004.599250] env[65121]: DEBUG oslo_concurrency.lockutils [req-bbdfd07f-bcb6-4621-b114-e1c00e6bd709 req-18e29b5e-1819-43bc-a29f-6944766951d6 service nova] Acquiring lock "efb98d37-4162-4249-9f85-008d4537db87-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.599472] env[65121]: DEBUG oslo_concurrency.lockutils [req-bbdfd07f-bcb6-4621-b114-e1c00e6bd709 req-18e29b5e-1819-43bc-a29f-6944766951d6 service nova] Lock "efb98d37-4162-4249-9f85-008d4537db87-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.599657] env[65121]: DEBUG oslo_concurrency.lockutils [req-bbdfd07f-bcb6-4621-b114-e1c00e6bd709 req-18e29b5e-1819-43bc-a29f-6944766951d6 service nova] Lock "efb98d37-4162-4249-9f85-008d4537db87-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.599853] env[65121]: DEBUG nova.compute.manager [req-bbdfd07f-bcb6-4621-b114-e1c00e6bd709 req-18e29b5e-1819-43bc-a29f-6944766951d6 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] No waiting events found dispatching network-vif-plugged-adb6945c-5a1e-487c-b116-4200874d80a0 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1004.600110] env[65121]: WARNING nova.compute.manager [req-bbdfd07f-bcb6-4621-b114-e1c00e6bd709 req-18e29b5e-1819-43bc-a29f-6944766951d6 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Received unexpected event network-vif-plugged-adb6945c-5a1e-487c-b116-4200874d80a0 for instance with vm_state building and task_state spawning. [ 1004.617424] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.617863] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.618227] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Deleting the datastore file [datastore1] 90cb98b1-9520-4caf-b6c4-80abaa7a9221 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.618627] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fbefafa-51f6-4df4-b12f-8486b63893b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.627446] env[65121]: DEBUG oslo_vmware.api [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for the task: (returnval){ [ 1004.627446] env[65121]: value = "task-5106962" [ 1004.627446] env[65121]: _type = "Task" [ 1004.627446] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.640897] env[65121]: DEBUG oslo_vmware.api [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106962, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.649542] env[65121]: DEBUG nova.network.neutron [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Successfully updated port: adb6945c-5a1e-487c-b116-4200874d80a0 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1004.707903] env[65121]: DEBUG nova.compute.utils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1004.712799] env[65121]: DEBUG nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1004.717252] env[65121]: DEBUG nova.network.neutron [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1004.717708] env[65121]: WARNING neutronclient.v2_0.client [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1004.718122] env[65121]: WARNING neutronclient.v2_0.client [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1004.718890] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.719358] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.732338] env[65121]: DEBUG nova.network.neutron [-] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.740679] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106960, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.754560] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b223eca7-6d42-4e99-a35c-03102bf90a1d tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.106s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.780677] env[65121]: DEBUG nova.policy [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26beb81d28f44cc8932ea8e87a4cd2c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0630960dcbf44781be05184565d81932', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1004.857898] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-524f9b67-267c-45d9-b38f-cb25d5f7a951 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.872144] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9e68d2-d042-45dd-964f-1b81a247a855 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.912428] env[65121]: DEBUG nova.compute.manager [req-0e37a9ef-746d-484a-a2a1-8573a3148129 req-c010d596-a037-4ab8-9901-5e61a0f6c008 service nova] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Detach interface failed, port_id=d9f74553-cfad-41e7-adf0-5faf0b793cf0, reason: Instance 17f34d32-e690-40d8-99a4-acf53fa859c6 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1005.109491] env[65121]: DEBUG nova.network.neutron [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Successfully created port: 3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1005.139390] env[65121]: DEBUG oslo_vmware.api [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Task: {'id': task-5106962, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147598} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.139663] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.139819] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1005.139994] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1005.140221] env[65121]: INFO nova.compute.manager [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1005.140409] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1005.140607] env[65121]: DEBUG nova.compute.manager [-] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1005.140727] env[65121]: DEBUG nova.network.neutron [-] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1005.141041] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.141589] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.141842] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.152039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "refresh_cache-efb98d37-4162-4249-9f85-008d4537db87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.152199] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "refresh_cache-efb98d37-4162-4249-9f85-008d4537db87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1005.152366] env[65121]: DEBUG nova.network.neutron [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1005.206181] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.213089] env[65121]: DEBUG nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1005.228484] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106960, 'name': Rename_Task, 'duration_secs': 0.874949} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.231414] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.231991] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf850de5-9cb0-4b6c-beeb-d31bfb31d9eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.234508] env[65121]: INFO nova.compute.manager [-] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Took 1.28 seconds to deallocate network for instance. [ 1005.241756] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1005.241756] env[65121]: value = "task-5106963" [ 1005.241756] env[65121]: _type = "Task" [ 1005.241756] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.251720] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.629059] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "84ae6061-a962-4a74-8661-9718cc4c5346" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.629386] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.634549] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320b8bc0-8eb1-49bf-9c08-7848f2695d09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.643346] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f5140a-57e6-404e-a5eb-ed308f564fa7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.676215] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.676618] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.686078] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51421507-50e1-49bc-8e70-8f7f3f6bd2ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.695088] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5389b9b9-fb6a-4bf9-8897-54c2a97ed145 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.722968] env[65121]: DEBUG nova.compute.provider_tree [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.748306] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.755432] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106963, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.809548] env[65121]: DEBUG nova.network.neutron [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1005.873132] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.873529] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.950376] env[65121]: WARNING neutronclient.v2_0.client [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.952861] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.952861] env[65121]: WARNING openstack [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.138635] env[65121]: DEBUG nova.network.neutron [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Updating instance_info_cache with network_info: [{"id": "adb6945c-5a1e-487c-b116-4200874d80a0", "address": "fa:16:3e:21:14:df", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadb6945c-5a", "ovs_interfaceid": "adb6945c-5a1e-487c-b116-4200874d80a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1006.194317] env[65121]: DEBUG nova.network.neutron [-] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1006.226955] env[65121]: DEBUG nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1006.234383] env[65121]: DEBUG nova.scheduler.client.report [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1006.260086] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106963, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.265618] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1006.265979] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1006.266722] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1006.266722] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1006.266722] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1006.266722] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1006.267101] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.267169] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1006.267378] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1006.267513] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1006.267682] env[65121]: DEBUG nova.virt.hardware [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1006.269198] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1960eb99-ddbb-45e0-8c4d-61dcc3e533da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.280230] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91885c0-f230-4d05-a50e-839ad6f37287 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.641414] env[65121]: DEBUG nova.compute.manager [req-2cedf05a-62a7-42c9-8722-16b71586af27 req-93a7bd2c-0578-40da-8b3b-05d8b82f51e3 service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Received event network-vif-plugged-3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1006.641694] env[65121]: DEBUG oslo_concurrency.lockutils [req-2cedf05a-62a7-42c9-8722-16b71586af27 req-93a7bd2c-0578-40da-8b3b-05d8b82f51e3 service nova] Acquiring lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.641763] env[65121]: DEBUG oslo_concurrency.lockutils [req-2cedf05a-62a7-42c9-8722-16b71586af27 req-93a7bd2c-0578-40da-8b3b-05d8b82f51e3 service nova] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.641926] env[65121]: DEBUG oslo_concurrency.lockutils [req-2cedf05a-62a7-42c9-8722-16b71586af27 req-93a7bd2c-0578-40da-8b3b-05d8b82f51e3 service nova] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1006.642197] env[65121]: DEBUG nova.compute.manager [req-2cedf05a-62a7-42c9-8722-16b71586af27 req-93a7bd2c-0578-40da-8b3b-05d8b82f51e3 service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] No waiting events found dispatching network-vif-plugged-3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1006.642270] env[65121]: WARNING nova.compute.manager [req-2cedf05a-62a7-42c9-8722-16b71586af27 req-93a7bd2c-0578-40da-8b3b-05d8b82f51e3 service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Received unexpected event network-vif-plugged-3bf01aba-25b1-44e7-a26e-8ffee646ba29 for instance with vm_state building and task_state spawning. [ 1006.643708] env[65121]: DEBUG nova.compute.manager [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Received event network-changed-adb6945c-5a1e-487c-b116-4200874d80a0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1006.643883] env[65121]: DEBUG nova.compute.manager [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Refreshing instance network info cache due to event network-changed-adb6945c-5a1e-487c-b116-4200874d80a0. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1006.644129] env[65121]: DEBUG oslo_concurrency.lockutils [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Acquiring lock "refresh_cache-efb98d37-4162-4249-9f85-008d4537db87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.644816] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "refresh_cache-efb98d37-4162-4249-9f85-008d4537db87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1006.645130] env[65121]: DEBUG nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Instance network_info: |[{"id": "adb6945c-5a1e-487c-b116-4200874d80a0", "address": "fa:16:3e:21:14:df", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadb6945c-5a", "ovs_interfaceid": "adb6945c-5a1e-487c-b116-4200874d80a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1006.645434] env[65121]: DEBUG oslo_concurrency.lockutils [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Acquired lock "refresh_cache-efb98d37-4162-4249-9f85-008d4537db87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1006.645696] env[65121]: DEBUG nova.network.neutron [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Refreshing network info cache for port adb6945c-5a1e-487c-b116-4200874d80a0 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1006.646811] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:14:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'adb6945c-5a1e-487c-b116-4200874d80a0', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1006.656242] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1006.657431] env[65121]: WARNING neutronclient.v2_0.client [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.658085] env[65121]: WARNING openstack [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.658430] env[65121]: WARNING openstack [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.665540] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efb98d37-4162-4249-9f85-008d4537db87] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1006.666248] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7cf5f958-f9ca-4052-81f0-db2cb02e2fdb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.688688] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1006.688688] env[65121]: value = "task-5106964" [ 1006.688688] env[65121]: _type = "Task" [ 1006.688688] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.698530] env[65121]: INFO nova.compute.manager [-] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Took 1.56 seconds to deallocate network for instance. [ 1006.698783] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106964, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.741452] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1006.742293] env[65121]: DEBUG nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1006.745148] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.997s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.747022] env[65121]: INFO nova.compute.claims [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.770033] env[65121]: DEBUG oslo_vmware.api [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5106963, 'name': PowerOnVM_Task, 'duration_secs': 1.07252} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.772412] env[65121]: DEBUG nova.network.neutron [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Successfully updated port: 3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1006.776885] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.776885] env[65121]: INFO nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Took 9.62 seconds to spawn the instance on the hypervisor. [ 1006.777050] env[65121]: DEBUG nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1006.778923] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12238088-cf6d-4f3f-98b2-0e79a797f8f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.791317] env[65121]: WARNING openstack [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.791722] env[65121]: WARNING openstack [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.858328] env[65121]: WARNING neutronclient.v2_0.client [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.859017] env[65121]: WARNING openstack [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.859373] env[65121]: WARNING openstack [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.946114] env[65121]: DEBUG nova.network.neutron [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Updated VIF entry in instance network info cache for port adb6945c-5a1e-487c-b116-4200874d80a0. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1006.946528] env[65121]: DEBUG nova.network.neutron [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Updating instance_info_cache with network_info: [{"id": "adb6945c-5a1e-487c-b116-4200874d80a0", "address": "fa:16:3e:21:14:df", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadb6945c-5a", "ovs_interfaceid": "adb6945c-5a1e-487c-b116-4200874d80a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1007.200608] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106964, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.205202] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.252521] env[65121]: DEBUG nova.compute.utils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1007.256659] env[65121]: DEBUG nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1007.278988] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.279156] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.279334] env[65121]: DEBUG nova.network.neutron [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1007.307698] env[65121]: INFO nova.compute.manager [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Took 50.72 seconds to build instance. [ 1007.450117] env[65121]: DEBUG oslo_concurrency.lockutils [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] Releasing lock "refresh_cache-efb98d37-4162-4249-9f85-008d4537db87" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1007.450370] env[65121]: DEBUG nova.compute.manager [req-6a59da1d-dc73-4f55-88b5-0893e180c143 req-8cd62831-28f6-438f-ab9e-e56e81ee2847 service nova] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Received event network-vif-deleted-86149f88-993f-45cf-94be-cb2781e74ac9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1007.700426] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106964, 'name': CreateVM_Task, 'duration_secs': 0.658345} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.700667] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efb98d37-4162-4249-9f85-008d4537db87] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1007.701200] env[65121]: WARNING neutronclient.v2_0.client [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.701748] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.701748] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.701996] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1007.702395] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dc3c5f3-3f0b-48e6-93f9-3cb76ba78d48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.707880] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1007.707880] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dc0ac6-12d7-9f7c-18c9-b30a29f2b2f4" [ 1007.707880] env[65121]: _type = "Task" [ 1007.707880] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.716903] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dc0ac6-12d7-9f7c-18c9-b30a29f2b2f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.757704] env[65121]: DEBUG nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1007.784723] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.785122] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.809870] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6524ee9a-5ed8-4cd6-b8af-d1d06572eb95 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.227s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.822805] env[65121]: DEBUG nova.network.neutron [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1007.844831] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.844929] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.924776] env[65121]: WARNING neutronclient.v2_0.client [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.925521] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.926224] env[65121]: WARNING openstack [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.017544] env[65121]: DEBUG nova.network.neutron [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Updating instance_info_cache with network_info: [{"id": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "address": "fa:16:3e:a8:19:39", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bf01aba-25", "ovs_interfaceid": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.150979] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66023df8-cf7d-4983-aa2a-dce7733eaea5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.160758] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0de22f4-67e9-4706-b603-5ba576bbce0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.193174] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418c9819-55ad-48fc-ac87-28017585d12e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.201265] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe71591-832a-47da-b3c5-32ed661020c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.216481] env[65121]: DEBUG nova.compute.provider_tree [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.226902] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dc0ac6-12d7-9f7c-18c9-b30a29f2b2f4, 'name': SearchDatastore_Task, 'duration_secs': 0.012668} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.227853] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.228100] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.228334] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.228471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1008.228641] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.229241] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23fe9c72-4893-48e3-84a2-05ddff998a50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.239058] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.239242] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1008.239944] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10090da9-8947-46e4-a783-832a4ded3d6d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.246141] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1008.246141] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52377516-113f-e851-e31f-09df10f7fe75" [ 1008.246141] env[65121]: _type = "Task" [ 1008.246141] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.254322] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52377516-113f-e851-e31f-09df10f7fe75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.317581] env[65121]: DEBUG nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1008.521218] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.521600] env[65121]: DEBUG nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Instance network_info: |[{"id": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "address": "fa:16:3e:a8:19:39", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bf01aba-25", "ovs_interfaceid": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1008.522081] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:19:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bf01aba-25b1-44e7-a26e-8ffee646ba29', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.529927] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1008.530156] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1008.530397] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2d2fc86-b1c8-40f9-a1d8-823bb89a0413 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.551362] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.551362] env[65121]: value = "task-5106965" [ 1008.551362] env[65121]: _type = "Task" [ 1008.551362] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.559942] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106965, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.684391] env[65121]: DEBUG nova.compute.manager [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Received event network-changed-3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1008.684860] env[65121]: DEBUG nova.compute.manager [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Refreshing instance network info cache due to event network-changed-3bf01aba-25b1-44e7-a26e-8ffee646ba29. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1008.684968] env[65121]: DEBUG oslo_concurrency.lockutils [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Acquiring lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.686136] env[65121]: DEBUG oslo_concurrency.lockutils [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Acquired lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1008.686136] env[65121]: DEBUG nova.network.neutron [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Refreshing network info cache for port 3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1008.722758] env[65121]: DEBUG nova.scheduler.client.report [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1008.757875] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52377516-113f-e851-e31f-09df10f7fe75, 'name': SearchDatastore_Task, 'duration_secs': 0.009774} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.758742] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa7a13e3-b57f-4588-9d6d-77259dd2bb3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.765668] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1008.765668] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523f7e0e-e7a0-e902-8431-80a20ac37635" [ 1008.765668] env[65121]: _type = "Task" [ 1008.765668] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.769972] env[65121]: DEBUG nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1008.779223] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523f7e0e-e7a0-e902-8431-80a20ac37635, 'name': SearchDatastore_Task, 'duration_secs': 0.011387} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.779449] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.779799] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] efb98d37-4162-4249-9f85-008d4537db87/efb98d37-4162-4249-9f85-008d4537db87.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1008.780130] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb4606f1-fe63-4ff1-a621-a750dad3c988 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.788629] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1008.788629] env[65121]: value = "task-5106966" [ 1008.788629] env[65121]: _type = "Task" [ 1008.788629] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.802017] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106966, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.804856] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1008.805159] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1008.805375] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1008.805644] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1008.805831] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1008.806048] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1008.806313] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.806427] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1008.806589] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1008.806749] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1008.806916] env[65121]: DEBUG nova.virt.hardware [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1008.807853] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94197e3-9959-4b66-99d3-7878daf78370 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.817366] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b64adf-44d6-443a-b980-f8d7c44ed38a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.839352] env[65121]: DEBUG nova.compute.manager [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1008.842448] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.849708] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Creating folder: Project (5a03645b81544bc58bcaa210a7184fa1). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1008.853060] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d157ed3a-7e3f-4e7f-8295-8062b5bc9d81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.866288] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Created folder: Project (5a03645b81544bc58bcaa210a7184fa1) in parent group-v993268. [ 1008.866657] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Creating folder: Instances. Parent ref: group-v993507. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1008.866966] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1ae97d6-9c82-4a04-a3a2-25e4aa7a695e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.870028] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.881828] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Created folder: Instances in parent group-v993507. [ 1008.882129] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1008.882352] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1008.882591] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-662325e9-bd60-4617-95d2-170d98d49e85 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.908444] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.908444] env[65121]: value = "task-5106969" [ 1008.908444] env[65121]: _type = "Task" [ 1008.908444] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.920575] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106969, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.062043] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106965, 'name': CreateVM_Task, 'duration_secs': 0.388935} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.062230] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.062777] env[65121]: WARNING neutronclient.v2_0.client [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.063209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.063368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1009.063700] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1009.063983] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cab400d4-0d54-48e5-b3ba-78f0b89c94ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.069768] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1009.069768] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52186931-553d-8023-a13f-5681d6b97104" [ 1009.069768] env[65121]: _type = "Task" [ 1009.069768] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.080086] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52186931-553d-8023-a13f-5681d6b97104, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.188980] env[65121]: WARNING neutronclient.v2_0.client [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.189927] env[65121]: WARNING openstack [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1009.190410] env[65121]: WARNING openstack [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1009.230360] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.231426] env[65121]: DEBUG nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1009.237091] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.574s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.237494] env[65121]: DEBUG nova.objects.instance [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lazy-loading 'resources' on Instance uuid 3f3bec5b-2834-497c-a454-a152cb992309 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.307861] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106966, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.358701] env[65121]: WARNING openstack [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1009.359108] env[65121]: WARNING openstack [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1009.371398] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.419419] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106969, 'name': CreateVM_Task, 'duration_secs': 0.374511} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.419654] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.420767] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.435931] env[65121]: WARNING neutronclient.v2_0.client [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.436629] env[65121]: WARNING openstack [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1009.437041] env[65121]: WARNING openstack [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1009.524119] env[65121]: DEBUG nova.network.neutron [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Updated VIF entry in instance network info cache for port 3bf01aba-25b1-44e7-a26e-8ffee646ba29. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1009.524525] env[65121]: DEBUG nova.network.neutron [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Updating instance_info_cache with network_info: [{"id": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "address": "fa:16:3e:a8:19:39", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bf01aba-25", "ovs_interfaceid": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1009.582213] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52186931-553d-8023-a13f-5681d6b97104, 'name': SearchDatastore_Task, 'duration_secs': 0.071775} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.582549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1009.582782] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.583034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.583181] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1009.583355] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.583632] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1009.583962] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1009.584216] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5582ae5-878d-4314-a2e6-40c54523b90e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.586247] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a53bfd9-9186-405f-a27b-1a048fef186e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.593178] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1009.593178] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5252158d-5b3f-e705-74b2-8b39d22e74da" [ 1009.593178] env[65121]: _type = "Task" [ 1009.593178] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.597810] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.597982] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.599127] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a93b1108-4bb4-4de6-b778-4bef49273854 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.604979] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5252158d-5b3f-e705-74b2-8b39d22e74da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.608421] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1009.608421] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52161b71-4eaf-1df6-a339-85708f620bfa" [ 1009.608421] env[65121]: _type = "Task" [ 1009.608421] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.616968] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52161b71-4eaf-1df6-a339-85708f620bfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.742523] env[65121]: DEBUG nova.compute.utils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1009.748736] env[65121]: DEBUG nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1009.748736] env[65121]: DEBUG nova.network.neutron [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1009.748945] env[65121]: WARNING neutronclient.v2_0.client [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.749260] env[65121]: WARNING neutronclient.v2_0.client [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.749883] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1009.750646] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1009.804582] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106966, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534775} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.805053] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] efb98d37-4162-4249-9f85-008d4537db87/efb98d37-4162-4249-9f85-008d4537db87.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1009.805280] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1009.805547] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d1a1a3e-ae16-4247-80fd-970ccff90030 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.811459] env[65121]: DEBUG nova.policy [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e773a3343a448f5acd3f08005e8ed45', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '40aee06073b04c25bc7b38fbd09856e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1009.821355] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1009.821355] env[65121]: value = "task-5106970" [ 1009.821355] env[65121]: _type = "Task" [ 1009.821355] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.841442] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.028310] env[65121]: DEBUG oslo_concurrency.lockutils [req-d5e14b04-0b14-44ec-9b14-78e379dc74b8 req-13672609-5516-4a49-86e7-e6c2e793494c service nova] Releasing lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1010.108662] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5252158d-5b3f-e705-74b2-8b39d22e74da, 'name': SearchDatastore_Task, 'duration_secs': 0.013087} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.111670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1010.111905] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.112174] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.122032] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52161b71-4eaf-1df6-a339-85708f620bfa, 'name': SearchDatastore_Task, 'duration_secs': 0.009804} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.122761] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c64b54de-ee7a-476e-8509-f680eaa8abc8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.129114] env[65121]: DEBUG nova.network.neutron [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Successfully created port: 62496d8b-5753-4794-b358-9db503a52328 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1010.135444] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1010.135444] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fcd835-1ca8-096f-ec60-82c8e0cb7b75" [ 1010.135444] env[65121]: _type = "Task" [ 1010.135444] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.144885] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fcd835-1ca8-096f-ec60-82c8e0cb7b75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.165281] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ad65cc-0559-45e4-9340-f6fde5d92939 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.174159] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4238e7-4540-41c6-b04d-2e43c3462a53 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.211619] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21391074-483a-4f87-b5b7-431a6b3fc6b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.220585] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e8eb8a-48ac-44a7-a57a-7a83cfaade59 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.242496] env[65121]: DEBUG nova.compute.provider_tree [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.249184] env[65121]: DEBUG nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1010.332985] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078383} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.333270] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1010.334237] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d604b6-9050-4c7b-aeca-d0bd0c2b967a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.360238] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] efb98d37-4162-4249-9f85-008d4537db87/efb98d37-4162-4249-9f85-008d4537db87.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.360958] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10a27825-c9e7-4932-9775-8c95a6edb90f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.382351] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1010.382351] env[65121]: value = "task-5106971" [ 1010.382351] env[65121]: _type = "Task" [ 1010.382351] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.391269] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.648447] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fcd835-1ca8-096f-ec60-82c8e0cb7b75, 'name': SearchDatastore_Task, 'duration_secs': 0.012602} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.648447] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1010.648447] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 6e969388-3238-404d-a8eb-e7b7318c4c72/6e969388-3238-404d-a8eb-e7b7318c4c72.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.648447] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1010.648447] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.648656] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-519a1a07-eead-4d76-a2db-53f8f4290691 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.650919] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d18b567d-e202-4c42-a4df-140adc3d4208 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.661257] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1010.661257] env[65121]: value = "task-5106972" [ 1010.661257] env[65121]: _type = "Task" [ 1010.661257] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.665826] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.666061] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1010.667295] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a21cf9eb-e71c-4a82-9a97-86045f46a83f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.673618] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.677881] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1010.677881] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524c9d77-bf3d-59c6-2c23-f9c0c45db59a" [ 1010.677881] env[65121]: _type = "Task" [ 1010.677881] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.689753] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524c9d77-bf3d-59c6-2c23-f9c0c45db59a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.746315] env[65121]: DEBUG nova.scheduler.client.report [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1010.895291] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106971, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.174201] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106972, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.191046] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524c9d77-bf3d-59c6-2c23-f9c0c45db59a, 'name': SearchDatastore_Task, 'duration_secs': 0.018408} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.191739] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89aa440d-2188-4c8a-bbcf-423efaa752a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.200917] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1011.200917] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520be31d-665e-edaf-76be-22ec7b1c2a01" [ 1011.200917] env[65121]: _type = "Task" [ 1011.200917] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.213360] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520be31d-665e-edaf-76be-22ec7b1c2a01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.252898] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.016s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.255987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 37.222s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.256392] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.256445] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1011.256719] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.175s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.257063] env[65121]: DEBUG nova.objects.instance [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lazy-loading 'resources' on Instance uuid 9cf6c29a-3424-4b88-9ba5-8120b124beb6 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.258944] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18c1f74-bda8-45c3-a215-fc417b10ef02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.263429] env[65121]: DEBUG nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1011.273075] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8912bf-fdf8-4af8-a3ff-92fe22917b7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.279615] env[65121]: INFO nova.scheduler.client.report [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted allocations for instance 3f3bec5b-2834-497c-a454-a152cb992309 [ 1011.300500] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c86fc3-ba9d-4f44-9385-faae5d92394c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.310040] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1011.310040] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1011.310258] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1011.310301] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1011.310439] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1011.310578] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1011.310781] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.310935] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1011.311110] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1011.311265] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1011.311427] env[65121]: DEBUG nova.virt.hardware [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1011.312554] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3eadc41-5929-443a-9724-7a5eeb19a06a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.317098] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04beebc-52ba-46ae-8323-fdb61a8d035e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.357486] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ce9297-4272-4a57-86f2-36f62dc9e4d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.360700] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178411MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1011.360855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.393922] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106971, 'name': ReconfigVM_Task, 'duration_secs': 0.680889} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.394237] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Reconfigured VM instance instance-00000055 to attach disk [datastore1] efb98d37-4162-4249-9f85-008d4537db87/efb98d37-4162-4249-9f85-008d4537db87.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1011.394948] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88c21629-018d-44ee-9d83-0a948d44d1af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.403115] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1011.403115] env[65121]: value = "task-5106973" [ 1011.403115] env[65121]: _type = "Task" [ 1011.403115] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.417410] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106973, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.618443] env[65121]: DEBUG nova.compute.manager [req-7b58f25a-401e-40a9-9a4b-7bcd4d68ed50 req-b96c2d98-a57e-4c15-8156-dd5ee9b968c9 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Received event network-vif-plugged-62496d8b-5753-4794-b358-9db503a52328 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1011.619276] env[65121]: DEBUG oslo_concurrency.lockutils [req-7b58f25a-401e-40a9-9a4b-7bcd4d68ed50 req-b96c2d98-a57e-4c15-8156-dd5ee9b968c9 service nova] Acquiring lock "e756fdb2-f901-4e23-92e7-f2753fd5f728-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.619276] env[65121]: DEBUG oslo_concurrency.lockutils [req-7b58f25a-401e-40a9-9a4b-7bcd4d68ed50 req-b96c2d98-a57e-4c15-8156-dd5ee9b968c9 service nova] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.619276] env[65121]: DEBUG oslo_concurrency.lockutils [req-7b58f25a-401e-40a9-9a4b-7bcd4d68ed50 req-b96c2d98-a57e-4c15-8156-dd5ee9b968c9 service nova] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.619570] env[65121]: DEBUG nova.compute.manager [req-7b58f25a-401e-40a9-9a4b-7bcd4d68ed50 req-b96c2d98-a57e-4c15-8156-dd5ee9b968c9 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] No waiting events found dispatching network-vif-plugged-62496d8b-5753-4794-b358-9db503a52328 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1011.619850] env[65121]: WARNING nova.compute.manager [req-7b58f25a-401e-40a9-9a4b-7bcd4d68ed50 req-b96c2d98-a57e-4c15-8156-dd5ee9b968c9 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Received unexpected event network-vif-plugged-62496d8b-5753-4794-b358-9db503a52328 for instance with vm_state building and task_state spawning. [ 1011.673527] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56974} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.674167] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 6e969388-3238-404d-a8eb-e7b7318c4c72/6e969388-3238-404d-a8eb-e7b7318c4c72.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.674451] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.674803] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2424c42b-3c36-47c7-bdfc-bfcbfbeb7943 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.685553] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1011.685553] env[65121]: value = "task-5106974" [ 1011.685553] env[65121]: _type = "Task" [ 1011.685553] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.694504] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106974, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.713652] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520be31d-665e-edaf-76be-22ec7b1c2a01, 'name': SearchDatastore_Task, 'duration_secs': 0.015108} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.714877] env[65121]: DEBUG nova.network.neutron [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Successfully updated port: 62496d8b-5753-4794-b358-9db503a52328 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1011.717160] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1011.718037] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.718864] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9bee281c-8640-45c2-b676-fd65adf03668 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.729533] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1011.729533] env[65121]: value = "task-5106975" [ 1011.729533] env[65121]: _type = "Task" [ 1011.729533] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.743582] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106975, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.805646] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0df4156d-7095-4258-ac32-0b329487991a tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "3f3bec5b-2834-497c-a454-a152cb992309" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.608s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.919087] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106973, 'name': Rename_Task, 'duration_secs': 0.445389} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.919370] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.919622] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4f2dae3-92bc-4c1a-afb0-93f4bf9ca13e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.929658] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1011.929658] env[65121]: value = "task-5106976" [ 1011.929658] env[65121]: _type = "Task" [ 1011.929658] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.939343] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.144490] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c43b391-68bd-48fc-9516-576a14346700 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.153197] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661fc3a7-2395-4b89-9aae-4297b4522444 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.190314] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2bf58d-9866-4e69-90eb-bd42e209e7ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.202644] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106974, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.251086} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.203281] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.204747] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae27cc5-5955-4b49-9f46-d68399fd59c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.209913] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d99a29-d9ca-4a0d-b31f-a01ceb52ab45 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.226281] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "refresh_cache-e756fdb2-f901-4e23-92e7-f2753fd5f728" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.226363] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "refresh_cache-e756fdb2-f901-4e23-92e7-f2753fd5f728" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1012.226590] env[65121]: DEBUG nova.network.neutron [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1012.228502] env[65121]: DEBUG nova.compute.provider_tree [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.251164] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 6e969388-3238-404d-a8eb-e7b7318c4c72/6e969388-3238-404d-a8eb-e7b7318c4c72.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.256495] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a9b5ace-67c2-4e35-9b24-cfcd800fede5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.279833] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106975, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.281822] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1012.281822] env[65121]: value = "task-5106977" [ 1012.281822] env[65121]: _type = "Task" [ 1012.281822] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.293576] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106977, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.442881] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106976, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.486686] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "274d0ccd-c707-4a68-b280-16de2bc74d73" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1012.487036] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.487517] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "274d0ccd-c707-4a68-b280-16de2bc74d73-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1012.487517] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.487667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1012.490090] env[65121]: INFO nova.compute.manager [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Terminating instance [ 1012.731577] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.731909] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.751033] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106975, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.810965} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.751312] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1012.751528] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.752494] env[65121]: DEBUG nova.scheduler.client.report [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.755765] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d4c9fcb-9d66-4699-8ea9-e62902099b0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.763206] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1012.763206] env[65121]: value = "task-5106978" [ 1012.763206] env[65121]: _type = "Task" [ 1012.763206] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.774159] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106978, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.775258] env[65121]: DEBUG nova.network.neutron [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1012.791356] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106977, 'name': ReconfigVM_Task, 'duration_secs': 0.490018} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.791666] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 6e969388-3238-404d-a8eb-e7b7318c4c72/6e969388-3238-404d-a8eb-e7b7318c4c72.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.792256] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4272ec8c-cc17-430a-8d94-b3028fcbd0b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.798077] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.798580] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.807471] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1012.807471] env[65121]: value = "task-5106979" [ 1012.807471] env[65121]: _type = "Task" [ 1012.807471] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.818337] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106979, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.875650] env[65121]: WARNING neutronclient.v2_0.client [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.876514] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.876850] env[65121]: WARNING openstack [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.941408] env[65121]: DEBUG oslo_vmware.api [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5106976, 'name': PowerOnVM_Task, 'duration_secs': 0.749922} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.941587] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1012.941792] env[65121]: INFO nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Took 8.77 seconds to spawn the instance on the hypervisor. [ 1012.941965] env[65121]: DEBUG nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1012.942925] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7883f2-ae36-4d5f-964a-9f6e744acd2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.972063] env[65121]: DEBUG nova.network.neutron [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Updating instance_info_cache with network_info: [{"id": "62496d8b-5753-4794-b358-9db503a52328", "address": "fa:16:3e:fd:e2:73", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62496d8b-57", "ovs_interfaceid": "62496d8b-5753-4794-b358-9db503a52328", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1012.996162] env[65121]: DEBUG nova.compute.manager [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1012.996162] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.996162] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d36cb3-fe4a-481e-b629-c7bea68d30df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.005999] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.006299] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6961b1d4-0b49-4eda-b954-f21af2a444f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.013638] env[65121]: DEBUG oslo_vmware.api [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1013.013638] env[65121]: value = "task-5106980" [ 1013.013638] env[65121]: _type = "Task" [ 1013.013638] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.022780] env[65121]: DEBUG oslo_vmware.api [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.260979] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.004s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.264606] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.509s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.267238] env[65121]: INFO nova.compute.claims [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.282572] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106978, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082862} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.282757] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.283528] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f02ffd-3109-46ce-b676-753470d01539 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.289301] env[65121]: INFO nova.scheduler.client.report [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted allocations for instance 9cf6c29a-3424-4b88-9ba5-8120b124beb6 [ 1013.307480] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.310348] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d945fb32-9375-48e7-a30e-b617af654506 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.338581] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106979, 'name': Rename_Task, 'duration_secs': 0.167141} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.340171] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.340432] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1013.340432] env[65121]: value = "task-5106981" [ 1013.340432] env[65121]: _type = "Task" [ 1013.340432] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.340637] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-829da47b-260f-4252-ac21-2fad6f110bae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.355575] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1013.355575] env[65121]: value = "task-5106982" [ 1013.355575] env[65121]: _type = "Task" [ 1013.355575] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.355802] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106981, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.368161] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.462194] env[65121]: INFO nova.compute.manager [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Took 51.89 seconds to build instance. [ 1013.475378] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "refresh_cache-e756fdb2-f901-4e23-92e7-f2753fd5f728" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1013.475538] env[65121]: DEBUG nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Instance network_info: |[{"id": "62496d8b-5753-4794-b358-9db503a52328", "address": "fa:16:3e:fd:e2:73", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62496d8b-57", "ovs_interfaceid": "62496d8b-5753-4794-b358-9db503a52328", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1013.476032] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:e2:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62496d8b-5753-4794-b358-9db503a52328', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.484197] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1013.485079] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.485345] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c261e957-cca8-4227-aef5-f06c45c4d569 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.507263] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.507263] env[65121]: value = "task-5106983" [ 1013.507263] env[65121]: _type = "Task" [ 1013.507263] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.517091] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106983, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.525158] env[65121]: DEBUG oslo_vmware.api [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106980, 'name': PowerOffVM_Task, 'duration_secs': 0.336731} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.525419] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1013.525569] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.525861] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba712fe8-3e5f-4cfc-98a2-59447dbd23e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.602391] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.602708] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.602995] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleting the datastore file [datastore1] 274d0ccd-c707-4a68-b280-16de2bc74d73 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.603340] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b836ce9-5991-42d5-88b4-0ec0efe55e60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.614163] env[65121]: DEBUG oslo_vmware.api [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1013.614163] env[65121]: value = "task-5106985" [ 1013.614163] env[65121]: _type = "Task" [ 1013.614163] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.626885] env[65121]: DEBUG oslo_vmware.api [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106985, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.663855] env[65121]: DEBUG nova.compute.manager [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Received event network-changed-62496d8b-5753-4794-b358-9db503a52328 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1013.664166] env[65121]: DEBUG nova.compute.manager [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Refreshing instance network info cache due to event network-changed-62496d8b-5753-4794-b358-9db503a52328. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1013.664434] env[65121]: DEBUG oslo_concurrency.lockutils [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Acquiring lock "refresh_cache-e756fdb2-f901-4e23-92e7-f2753fd5f728" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.664578] env[65121]: DEBUG oslo_concurrency.lockutils [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Acquired lock "refresh_cache-e756fdb2-f901-4e23-92e7-f2753fd5f728" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.664734] env[65121]: DEBUG nova.network.neutron [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Refreshing network info cache for port 62496d8b-5753-4794-b358-9db503a52328 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1013.830854] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f058c255-5059-4e25-b5d1-36fa4c961348 tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "9cf6c29a-3424-4b88-9ba5-8120b124beb6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.187s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.855157] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106981, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.867322] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106982, 'name': PowerOnVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.964792] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f7a159c5-a8b9-45f3-b2aa-e80d88eeb86a tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "efb98d37-4162-4249-9f85-008d4537db87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.423s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.021013] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106983, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.125570] env[65121]: DEBUG oslo_vmware.api [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5106985, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19633} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.125847] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.126083] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.126287] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.126458] env[65121]: INFO nova.compute.manager [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1014.127215] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1014.127215] env[65121]: DEBUG nova.compute.manager [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1014.127215] env[65121]: DEBUG nova.network.neutron [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1014.127483] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.128221] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.128491] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.168206] env[65121]: WARNING neutronclient.v2_0.client [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.169455] env[65121]: WARNING openstack [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.171076] env[65121]: WARNING openstack [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.190723] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.335990] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "3d6fc9df-6872-4929-8813-bec1db1ebb16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.336481] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.357152] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106981, 'name': ReconfigVM_Task, 'duration_secs': 0.801366} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.357536] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Reconfigured VM instance instance-00000057 to attach disk [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.361506] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0cd61d8-14f8-4ab9-9bf1-5562ed422c10 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.370241] env[65121]: DEBUG oslo_vmware.api [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106982, 'name': PowerOnVM_Task, 'duration_secs': 0.700317} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.374701] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.374950] env[65121]: INFO nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Took 8.15 seconds to spawn the instance on the hypervisor. [ 1014.376068] env[65121]: DEBUG nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1014.376068] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1014.376068] env[65121]: value = "task-5106986" [ 1014.376068] env[65121]: _type = "Task" [ 1014.376068] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.377556] env[65121]: WARNING openstack [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.377917] env[65121]: WARNING openstack [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.385490] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3690bf-7896-4600-94cc-ccc7572f770d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.409129] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106986, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.469138] env[65121]: DEBUG nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1014.483478] env[65121]: WARNING neutronclient.v2_0.client [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.484203] env[65121]: WARNING openstack [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.484534] env[65121]: WARNING openstack [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.519809] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5106983, 'name': CreateVM_Task, 'duration_secs': 0.735811} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.522539] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.523662] env[65121]: WARNING neutronclient.v2_0.client [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.523662] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.524026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.524091] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1014.524666] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ecabfec-e785-4a37-bab8-2b9f73526fc9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.530938] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1014.530938] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52aa8c14-9b82-86d1-a7eb-caa2bce4a9da" [ 1014.530938] env[65121]: _type = "Task" [ 1014.530938] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.553202] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52aa8c14-9b82-86d1-a7eb-caa2bce4a9da, 'name': SearchDatastore_Task, 'duration_secs': 0.014735} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.553368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.553604] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.553833] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.554045] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.554521] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.554691] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2e4964f-17b6-493e-a02b-80c6db213f87 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.566020] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.566158] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.566950] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0504c0f4-a6ae-47f2-b9bb-c1190e488d30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.574262] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1014.574262] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a04f3b-b3d6-d04f-33ad-92d032467651" [ 1014.574262] env[65121]: _type = "Task" [ 1014.574262] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.590238] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a04f3b-b3d6-d04f-33ad-92d032467651, 'name': SearchDatastore_Task, 'duration_secs': 0.013741} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.591604] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de10b6d0-6dd8-4d22-984c-5879fd10c4db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.601016] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1014.601016] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c9c480-00af-3675-f79e-6cdc2597becc" [ 1014.601016] env[65121]: _type = "Task" [ 1014.601016] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.611212] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c9c480-00af-3675-f79e-6cdc2597becc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.728052] env[65121]: DEBUG nova.network.neutron [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Updated VIF entry in instance network info cache for port 62496d8b-5753-4794-b358-9db503a52328. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1014.728420] env[65121]: DEBUG nova.network.neutron [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Updating instance_info_cache with network_info: [{"id": "62496d8b-5753-4794-b358-9db503a52328", "address": "fa:16:3e:fd:e2:73", "network": {"id": "fb1397df-ab8d-4ded-83f4-6d062c3e4d4c", "bridge": "br-int", "label": "tempest-ImagesTestJSON-18525116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "40aee06073b04c25bc7b38fbd09856e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62496d8b-57", "ovs_interfaceid": "62496d8b-5753-4794-b358-9db503a52328", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1014.765631] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a784d7-f625-4575-8cc6-35807733cfe3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.774463] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e968a48-2f33-45cb-a43e-b0e9ebc45c8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.816813] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c56990-c391-4448-afd0-c9cc131336e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.826879] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88689cd-58b8-4c10-b065-43443552d445 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.841856] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "27940143-16b5-4263-b23c-354ed8ea8866" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.842154] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "27940143-16b5-4263-b23c-354ed8ea8866" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.842303] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "27940143-16b5-4263-b23c-354ed8ea8866-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.842476] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "27940143-16b5-4263-b23c-354ed8ea8866-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.842993] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "27940143-16b5-4263-b23c-354ed8ea8866-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.844665] env[65121]: DEBUG nova.compute.provider_tree [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.846175] env[65121]: INFO nova.compute.manager [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Terminating instance [ 1014.901958] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106986, 'name': Rename_Task, 'duration_secs': 0.329515} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.902374] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.902645] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-148c34b9-31e7-455e-aa4c-c37c0c9d70a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.916073] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1014.916073] env[65121]: value = "task-5106987" [ 1014.916073] env[65121]: _type = "Task" [ 1014.916073] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.921127] env[65121]: INFO nova.compute.manager [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Took 48.70 seconds to build instance. [ 1014.928242] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106987, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.995550] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.112951] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c9c480-00af-3675-f79e-6cdc2597becc, 'name': SearchDatastore_Task, 'duration_secs': 0.015862} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.113278] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.113365] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e756fdb2-f901-4e23-92e7-f2753fd5f728/e756fdb2-f901-4e23-92e7-f2753fd5f728.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.113624] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45a90487-154c-4fb2-a5e8-592b371a3e3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.121336] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1015.121336] env[65121]: value = "task-5106988" [ 1015.121336] env[65121]: _type = "Task" [ 1015.121336] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.130920] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.231550] env[65121]: DEBUG oslo_concurrency.lockutils [req-ce9376c7-61af-408f-9748-2e0779f7f740 req-31dd976b-4eff-480a-9d6e-0e84cee88c90 service nova] Releasing lock "refresh_cache-e756fdb2-f901-4e23-92e7-f2753fd5f728" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.321327] env[65121]: DEBUG nova.network.neutron [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.349371] env[65121]: DEBUG nova.scheduler.client.report [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1015.353696] env[65121]: DEBUG nova.compute.manager [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1015.354022] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.355178] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6ab974-b237-4830-a94f-863101d8ce8b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.365392] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.365703] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d59a628b-5bea-4a99-8063-1c0f04653671 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.374125] env[65121]: DEBUG oslo_vmware.api [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 1015.374125] env[65121]: value = "task-5106989" [ 1015.374125] env[65121]: _type = "Task" [ 1015.374125] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.386104] env[65121]: DEBUG oslo_vmware.api [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106989, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.423225] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee73cc08-f7b1-44c4-873b-f9759c71c62e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.209s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.432716] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106987, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.633318] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106988, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.692580] env[65121]: DEBUG nova.compute.manager [req-42a8f80e-cef3-4dfe-b50a-8f53000476c2 req-cea21b02-fd09-4eaa-91f6-a885abbd5adb service nova] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Received event network-vif-deleted-748bfc22-93f1-459b-9b59-18583587dd17 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1015.824538] env[65121]: INFO nova.compute.manager [-] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Took 1.70 seconds to deallocate network for instance. [ 1015.855504] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.856460] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1015.859416] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.050s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.860839] env[65121]: INFO nova.compute.claims [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.887127] env[65121]: DEBUG oslo_vmware.api [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106989, 'name': PowerOffVM_Task, 'duration_secs': 0.254181} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.887568] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.887890] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.888225] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81548fe6-c573-4111-a531-1b3b70028ae5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.927964] env[65121]: DEBUG nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1015.934403] env[65121]: DEBUG oslo_vmware.api [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106987, 'name': PowerOnVM_Task, 'duration_secs': 0.744033} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.935333] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.935333] env[65121]: INFO nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Took 7.16 seconds to spawn the instance on the hypervisor. [ 1015.935333] env[65121]: DEBUG nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1015.936138] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65c77cd-2954-4f52-a356-daba1fac94a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.971411] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.971651] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.971837] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleting the datastore file [datastore2] 27940143-16b5-4263-b23c-354ed8ea8866 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.973077] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37801d87-5de1-4d10-ab53-272f46beab07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.984052] env[65121]: DEBUG oslo_vmware.api [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for the task: (returnval){ [ 1015.984052] env[65121]: value = "task-5106991" [ 1015.984052] env[65121]: _type = "Task" [ 1015.984052] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.996433] env[65121]: DEBUG oslo_vmware.api [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.133267] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.665713} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.133542] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e756fdb2-f901-4e23-92e7-f2753fd5f728/e756fdb2-f901-4e23-92e7-f2753fd5f728.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.133754] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.134049] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-adcd213e-2ea6-4c33-be05-2136d6658dcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.141692] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1016.141692] env[65121]: value = "task-5106992" [ 1016.141692] env[65121]: _type = "Task" [ 1016.141692] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.150983] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106992, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.276545] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "6e969388-3238-404d-a8eb-e7b7318c4c72" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.276796] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.277034] env[65121]: INFO nova.compute.manager [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Shelving [ 1016.332618] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.368208] env[65121]: DEBUG nova.compute.utils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1016.369623] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1016.369812] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1016.370140] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.370448] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.371019] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.371362] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.426377] env[65121]: DEBUG nova.policy [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '22de4b1af65e4ad88eff687a4fe94204', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4feb04f47a97427c8a2bc97883332e1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1016.455249] env[65121]: INFO nova.compute.manager [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Took 46.95 seconds to build instance. [ 1016.457203] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.494749] env[65121]: DEBUG oslo_vmware.api [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Task: {'id': task-5106991, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.51085} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.495025] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.495208] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.495381] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.495556] env[65121]: INFO nova.compute.manager [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1016.495821] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1016.496051] env[65121]: DEBUG nova.compute.manager [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1016.496153] env[65121]: DEBUG nova.network.neutron [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1016.496399] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.496936] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.497218] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.598132] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.652465] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106992, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117208} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.653554] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.653670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111c30b8-a04a-4e30-b69e-f05fc72a8d3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.678046] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] e756fdb2-f901-4e23-92e7-f2753fd5f728/e756fdb2-f901-4e23-92e7-f2753fd5f728.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.678215] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a780ecdd-ddf5-47a7-866e-46691fce6097 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.701946] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1016.701946] env[65121]: value = "task-5106993" [ 1016.701946] env[65121]: _type = "Task" [ 1016.701946] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.715260] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.758186] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Successfully created port: acee4131-75b4-4eef-9802-d42c3806fa26 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1016.879742] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1016.958309] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62264a05-d8f6-453a-80e5-01a2757f7963 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.462s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.213453] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106993, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.288048] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1017.288428] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-176a9778-7913-4ced-8691-b2add96c87b0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.298144] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1017.298144] env[65121]: value = "task-5106994" [ 1017.298144] env[65121]: _type = "Task" [ 1017.298144] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.309223] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.362335] env[65121]: DEBUG nova.network.neutron [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.371786] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6972ce20-36ad-4eb7-a9b4-e6c7aca6692f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.384951] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eade364a-458e-4177-90f8-058351a78db6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.428388] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4784c695-9f9d-4b2c-86e2-11a9ab57ce56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.438307] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deabab98-a1c5-4132-8c62-eefeba60e540 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.456462] env[65121]: DEBUG nova.compute.provider_tree [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.717061] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106993, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.740222] env[65121]: DEBUG nova.compute.manager [req-51e820bd-87f4-4b34-91de-89a12246ccaf req-674cdd4f-d075-4c85-91cf-6e3ac68bf408 service nova] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Received event network-vif-deleted-c2770914-538c-4cb2-9133-0998f766b7ad {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1017.810159] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106994, 'name': PowerOffVM_Task, 'duration_secs': 0.25528} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.810515] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1017.811608] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c69739-bd62-4272-a0bf-ca25bab9ce44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.832644] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f32d50-9a2b-44be-81f5-989976081604 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.868168] env[65121]: INFO nova.compute.manager [-] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Took 1.37 seconds to deallocate network for instance. [ 1017.868727] env[65121]: INFO nova.compute.manager [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Rebuilding instance [ 1017.895520] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1017.921139] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1017.921425] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1017.921567] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1017.921719] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1017.921858] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1017.921996] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1017.922212] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.922361] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1017.922518] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1017.922669] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1017.922828] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1017.924233] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efbd07f-40f9-4038-85cb-6e26cdf0d04b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.927682] env[65121]: DEBUG nova.compute.manager [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1017.928463] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f43f8ca-0eb7-469c-ae8b-48822a49ade1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.941513] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65458ef4-1d99-402f-b859-95726ba0f9c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.960539] env[65121]: DEBUG nova.scheduler.client.report [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1018.153989] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.153989] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.214058] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106993, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.345153] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1018.345551] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-192b46cb-fc7a-4d98-b387-aa8a53444311 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.356321] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1018.356321] env[65121]: value = "task-5106995" [ 1018.356321] env[65121]: _type = "Task" [ 1018.356321] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.366218] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106995, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.378558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.465763] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Successfully updated port: acee4131-75b4-4eef-9802-d42c3806fa26 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1018.468256] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.468781] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1018.471577] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.632s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.473044] env[65121]: INFO nova.compute.claims [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1018.520683] env[65121]: DEBUG nova.compute.manager [req-d533e5df-324a-4e78-9bcf-7874ae54004e req-c39d3db6-b583-4607-b199-9565748035ce service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Received event network-vif-plugged-acee4131-75b4-4eef-9802-d42c3806fa26 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1018.521012] env[65121]: DEBUG oslo_concurrency.lockutils [req-d533e5df-324a-4e78-9bcf-7874ae54004e req-c39d3db6-b583-4607-b199-9565748035ce service nova] Acquiring lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.521426] env[65121]: DEBUG oslo_concurrency.lockutils [req-d533e5df-324a-4e78-9bcf-7874ae54004e req-c39d3db6-b583-4607-b199-9565748035ce service nova] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.521781] env[65121]: DEBUG oslo_concurrency.lockutils [req-d533e5df-324a-4e78-9bcf-7874ae54004e req-c39d3db6-b583-4607-b199-9565748035ce service nova] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1018.522133] env[65121]: DEBUG nova.compute.manager [req-d533e5df-324a-4e78-9bcf-7874ae54004e req-c39d3db6-b583-4607-b199-9565748035ce service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] No waiting events found dispatching network-vif-plugged-acee4131-75b4-4eef-9802-d42c3806fa26 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1018.522419] env[65121]: WARNING nova.compute.manager [req-d533e5df-324a-4e78-9bcf-7874ae54004e req-c39d3db6-b583-4607-b199-9565748035ce service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Received unexpected event network-vif-plugged-acee4131-75b4-4eef-9802-d42c3806fa26 for instance with vm_state building and task_state spawning. [ 1018.585773] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.586033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.657666] env[65121]: DEBUG nova.compute.utils [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1018.713955] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106993, 'name': ReconfigVM_Task, 'duration_secs': 1.646571} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.714318] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Reconfigured VM instance instance-00000058 to attach disk [datastore1] e756fdb2-f901-4e23-92e7-f2753fd5f728/e756fdb2-f901-4e23-92e7-f2753fd5f728.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.714945] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d264e63c-399e-45c3-8e34-7394c2b52e67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.723154] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1018.723154] env[65121]: value = "task-5106996" [ 1018.723154] env[65121]: _type = "Task" [ 1018.723154] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.732256] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106996, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.866987] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106995, 'name': CreateSnapshot_Task, 'duration_secs': 0.42195} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.867341] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1018.868119] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e50046-416a-4cc2-89b0-79ac691b1fff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.945842] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.946192] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53251fc3-4a90-4665-9f8d-adfb30ddf649 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.953897] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1018.953897] env[65121]: value = "task-5106997" [ 1018.953897] env[65121]: _type = "Task" [ 1018.953897] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.962764] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.972745] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "refresh_cache-4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.972987] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "refresh_cache-4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.973205] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1018.977479] env[65121]: DEBUG nova.compute.utils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1018.981848] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1018.982230] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1018.982695] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.983087] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.983880] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.984332] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.038686] env[65121]: DEBUG nova.policy [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '22de4b1af65e4ad88eff687a4fe94204', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4feb04f47a97427c8a2bc97883332e1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1019.089625] env[65121]: DEBUG nova.compute.utils [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1019.161253] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.233961] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106996, 'name': Rename_Task, 'duration_secs': 0.294538} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.234271] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.234526] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b003df2b-02a9-4028-90e1-8736ce5330ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.241829] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1019.241829] env[65121]: value = "task-5106998" [ 1019.241829] env[65121]: _type = "Task" [ 1019.241829] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.250381] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.378443] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Successfully created port: 84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1019.387988] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1019.389159] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c6d66636-5d14-4919-9e39-89cc20683491 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.397580] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1019.397580] env[65121]: value = "task-5106999" [ 1019.397580] env[65121]: _type = "Task" [ 1019.397580] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.407042] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106999, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.463527] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5106997, 'name': PowerOffVM_Task, 'duration_secs': 0.214534} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.463773] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.464045] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1019.464804] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69db58f-4712-4693-80d0-c0b70cfa4c81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.472813] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1019.473629] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-417be240-a443-4415-9a69-bc1baa6edabe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.476171] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.476445] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.484485] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1019.523184] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1019.528041] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1019.528255] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1019.528521] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Deleting the datastore file [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.528702] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c055424-db6f-4115-aa83-42283cd56f62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.536827] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1019.536827] env[65121]: value = "task-5107001" [ 1019.536827] env[65121]: _type = "Task" [ 1019.536827] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.548281] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.548716] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.561499] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.593397] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.622079] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.623114] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.623650] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.728846] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Updating instance_info_cache with network_info: [{"id": "acee4131-75b4-4eef-9802-d42c3806fa26", "address": "fa:16:3e:e5:08:fb", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacee4131-75", "ovs_interfaceid": "acee4131-75b4-4eef-9802-d42c3806fa26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.759920] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106998, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.912434] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106999, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.931402] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc93b888-77f1-4321-b6a1-4279ca0fd700 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.940075] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1799e29f-cfa7-4dda-801e-0be7356c31e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.974498] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fd9cfd-00e1-4da1-abd2-da4f3d904acb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.986385] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0e4ab0-a042-4cbe-a222-3ca4cd6db5f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.005669] env[65121]: DEBUG nova.compute.provider_tree [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.047444] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107001, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113336} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.047708] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.047882] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1020.048069] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1020.235286] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "refresh_cache-4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.235717] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Instance network_info: |[{"id": "acee4131-75b4-4eef-9802-d42c3806fa26", "address": "fa:16:3e:e5:08:fb", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacee4131-75", "ovs_interfaceid": "acee4131-75b4-4eef-9802-d42c3806fa26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1020.236319] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:08:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eed34ae1-5f7f-4deb-9db8-85eaa1e60c29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'acee4131-75b4-4eef-9802-d42c3806fa26', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.245239] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Creating folder: Project (4feb04f47a97427c8a2bc97883332e1d). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1020.245734] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1020.245993] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1020.246280] env[65121]: INFO nova.compute.manager [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Attaching volume 7bcf71f4-0c34-441a-9533-931791d14ff6 to /dev/sdb [ 1020.248497] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f8faeb0-e22e-4e50-94c5-8a76045cfe3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.262105] env[65121]: DEBUG oslo_vmware.api [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5106998, 'name': PowerOnVM_Task, 'duration_secs': 0.676657} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.262379] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.262580] env[65121]: INFO nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Took 9.00 seconds to spawn the instance on the hypervisor. [ 1020.262797] env[65121]: DEBUG nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1020.264692] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4300268-9a15-4a4b-8fb5-05abfcc73a08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.267334] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Created folder: Project (4feb04f47a97427c8a2bc97883332e1d) in parent group-v993268. [ 1020.267532] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Creating folder: Instances. Parent ref: group-v993513. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1020.267790] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20de5b2b-04cc-4ae5-a091-0388559e0ce3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.280742] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Created folder: Instances in parent group-v993513. [ 1020.280742] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1020.280742] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.280742] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31c54181-dd96-4003-9a3c-fc622af23d18 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.298170] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72634b4c-24a1-4517-8196-ef5ce4353e1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.309032] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe655ff-c946-4572-a74d-04670b85bcf5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.312030] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.312030] env[65121]: value = "task-5107004" [ 1020.312030] env[65121]: _type = "Task" [ 1020.312030] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.324041] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107004, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.329581] env[65121]: DEBUG nova.virt.block_device [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Updating existing volume attachment record: eb90c0b7-b5dc-44b8-aa29-42df5cf2d3e4 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1020.411640] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106999, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.496754] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1020.509728] env[65121]: DEBUG nova.scheduler.client.report [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1020.529782] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1020.529952] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1020.530132] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1020.530314] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1020.530452] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1020.530592] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1020.530790] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.530937] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1020.531187] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1020.531280] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1020.531420] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1020.532561] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260d5f7e-f6fd-4b66-b39a-6d60d659772f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.542443] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4531ba2-e27f-428c-946f-cfd0714874b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.654516] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1020.654809] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1020.655052] env[65121]: INFO nova.compute.manager [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Attaching volume 79bb9f39-f959-4b03-8058-0e7401b28853 to /dev/sdb [ 1020.699404] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f09bbb-b8cb-432d-878f-f2030b8dcd5b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.705761] env[65121]: DEBUG nova.compute.manager [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Received event network-changed-acee4131-75b4-4eef-9802-d42c3806fa26 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1020.705971] env[65121]: DEBUG nova.compute.manager [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Refreshing instance network info cache due to event network-changed-acee4131-75b4-4eef-9802-d42c3806fa26. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1020.706318] env[65121]: DEBUG oslo_concurrency.lockutils [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Acquiring lock "refresh_cache-4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.706468] env[65121]: DEBUG oslo_concurrency.lockutils [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Acquired lock "refresh_cache-4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.706857] env[65121]: DEBUG nova.network.neutron [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Refreshing network info cache for port acee4131-75b4-4eef-9802-d42c3806fa26 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1020.714600] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7f1099-7c13-435d-8651-b4037baeab25 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.730816] env[65121]: DEBUG nova.virt.block_device [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating existing volume attachment record: 090ec1d1-9cde-4480-a02a-31985b637142 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1020.788965] env[65121]: INFO nova.compute.manager [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Took 49.06 seconds to build instance. [ 1020.823893] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107004, 'name': CreateVM_Task, 'duration_secs': 0.406618} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.824228] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.824784] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.825176] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.825327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.825647] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1020.826302] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecd1cf0b-9b3c-4cff-a19d-6d488017ab36 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.832067] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1020.832067] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52336f23-701c-5f71-bd95-da3485a046e3" [ 1020.832067] env[65121]: _type = "Task" [ 1020.832067] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.840929] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52336f23-701c-5f71-bd95-da3485a046e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.912187] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5106999, 'name': CloneVM_Task, 'duration_secs': 1.225914} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.912601] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Created linked-clone VM from snapshot [ 1020.913512] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de06aee-1d9e-4c44-9b2e-5f43c10456ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.922444] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Uploading image 3cc49a60-ee19-44c9-8abb-844dc06b8184 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1020.947068] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Successfully updated port: 84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1020.966077] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1020.966077] env[65121]: value = "vm-993512" [ 1020.966077] env[65121]: _type = "VirtualMachine" [ 1020.966077] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1020.966803] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3f252fa4-b40f-4956-9d29-2acf7fe10caa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.978047] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lease: (returnval){ [ 1020.978047] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523fc22b-dc1e-4ef5-d764-de1005d5e733" [ 1020.978047] env[65121]: _type = "HttpNfcLease" [ 1020.978047] env[65121]: } obtained for exporting VM: (result){ [ 1020.978047] env[65121]: value = "vm-993512" [ 1020.978047] env[65121]: _type = "VirtualMachine" [ 1020.978047] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1020.978624] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the lease: (returnval){ [ 1020.978624] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523fc22b-dc1e-4ef5-d764-de1005d5e733" [ 1020.978624] env[65121]: _type = "HttpNfcLease" [ 1020.978624] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1020.986359] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1020.986359] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523fc22b-dc1e-4ef5-d764-de1005d5e733" [ 1020.986359] env[65121]: _type = "HttpNfcLease" [ 1020.986359] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1021.017068] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.017729] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1021.020765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.999s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.022447] env[65121]: INFO nova.compute.claims [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.090617] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1021.090860] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1021.091024] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1021.091201] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1021.091341] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1021.091480] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1021.091680] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.091856] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1021.091983] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1021.092161] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1021.092327] env[65121]: DEBUG nova.virt.hardware [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1021.093519] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a58ee9-068e-42e7-a652-a225b470bb19 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.101351] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34a8e28-dcb3-4048-bcc2-5fede02be8cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.117337] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1021.123879] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1021.124433] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1021.124609] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cfc54e7-e7f2-4a48-a157-6e3e43fda1a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.143727] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1021.143727] env[65121]: value = "task-5107010" [ 1021.143727] env[65121]: _type = "Task" [ 1021.143727] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.152911] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107010, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.210828] env[65121]: WARNING neutronclient.v2_0.client [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1021.211879] env[65121]: WARNING openstack [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.212468] env[65121]: WARNING openstack [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.291669] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e03615af-e645-42ad-b356-4f6f5610e471 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.570s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.351075] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52336f23-701c-5f71-bd95-da3485a046e3, 'name': SearchDatastore_Task, 'duration_secs': 0.011275} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.351075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.351670] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.351670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.351810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.352070] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.352557] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2229165-1178-481e-bd05-e9330609fca4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.364121] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.364554] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.367118] env[65121]: WARNING openstack [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.367793] env[65121]: WARNING openstack [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.376134] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8c8574f-da3b-4432-ab78-d519ed58e32c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.383110] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1021.383110] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52be5b62-85a4-17c7-6607-efd2b8cfd966" [ 1021.383110] env[65121]: _type = "Task" [ 1021.383110] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.394267] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be5b62-85a4-17c7-6607-efd2b8cfd966, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.452707] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "refresh_cache-ec463cad-4c80-4636-bc7a-9ec298a07d96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.453106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "refresh_cache-ec463cad-4c80-4636-bc7a-9ec298a07d96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.453106] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1021.455763] env[65121]: WARNING neutronclient.v2_0.client [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1021.456542] env[65121]: WARNING openstack [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.456807] env[65121]: WARNING openstack [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.486722] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1021.486722] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523fc22b-dc1e-4ef5-d764-de1005d5e733" [ 1021.486722] env[65121]: _type = "HttpNfcLease" [ 1021.486722] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1021.487102] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1021.487102] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523fc22b-dc1e-4ef5-d764-de1005d5e733" [ 1021.487102] env[65121]: _type = "HttpNfcLease" [ 1021.487102] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1021.487772] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8df0fa-9bf0-4ef7-9ae3-51cdffcc037c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.496223] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271e88b-434f-414f-2a98-2cff23d7bfe9/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1021.496503] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271e88b-434f-414f-2a98-2cff23d7bfe9/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1021.571604] env[65121]: DEBUG nova.compute.utils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1021.581799] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1021.582056] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1021.582404] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1021.582698] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1021.583300] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.583646] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.621077] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9e61f931-0eef-4077-84c2-a9a1d1faae55 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.635170] env[65121]: DEBUG nova.network.neutron [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Updated VIF entry in instance network info cache for port acee4131-75b4-4eef-9802-d42c3806fa26. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1021.639025] env[65121]: DEBUG nova.network.neutron [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Updating instance_info_cache with network_info: [{"id": "acee4131-75b4-4eef-9802-d42c3806fa26", "address": "fa:16:3e:e5:08:fb", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacee4131-75", "ovs_interfaceid": "acee4131-75b4-4eef-9802-d42c3806fa26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1021.641131] env[65121]: DEBUG nova.policy [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '22de4b1af65e4ad88eff687a4fe94204', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4feb04f47a97427c8a2bc97883332e1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1021.662181] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107010, 'name': CreateVM_Task, 'duration_secs': 0.305223} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.668609] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1021.670471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.670471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.670471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1021.670471] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05304811-f7d4-4398-a344-0ba794e2fc91 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.675849] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1021.675849] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52650f73-322b-04c0-f1ff-f29abbef7875" [ 1021.675849] env[65121]: _type = "Task" [ 1021.675849] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.685309] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52650f73-322b-04c0-f1ff-f29abbef7875, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.896684] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52be5b62-85a4-17c7-6607-efd2b8cfd966, 'name': SearchDatastore_Task, 'duration_secs': 0.011027} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.897757] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8afa6f4f-3a05-4893-93ae-2a79a06bd458 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.904748] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1021.904748] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525513e7-00fb-36c8-aa1b-497b39867995" [ 1021.904748] env[65121]: _type = "Task" [ 1021.904748] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.914174] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525513e7-00fb-36c8-aa1b-497b39867995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.956424] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Successfully created port: 6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1021.959357] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.959780] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.002156] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1022.026118] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.026735] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.096078] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1022.110545] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1022.111011] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.111622] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.139516] env[65121]: DEBUG oslo_concurrency.lockutils [req-16e77c87-e639-495c-903f-9ef6620ca1f5 req-15514dbd-f4d6-419e-b126-44dffac5134f service nova] Releasing lock "refresh_cache-4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.188332] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52650f73-322b-04c0-f1ff-f29abbef7875, 'name': SearchDatastore_Task, 'duration_secs': 0.010957} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.191729] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.192433] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.192433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.192433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.192739] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.193501] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-423e3d85-4966-4d41-859a-d6fa3b2e50a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.204591] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.204866] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1022.205945] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0a2548a-db28-4cf5-9466-bd275f7f425f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.216598] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1022.216598] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520e9a45-5092-128e-fc17-7679edf8a488" [ 1022.216598] env[65121]: _type = "Task" [ 1022.216598] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.228340] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Updating instance_info_cache with network_info: [{"id": "84a9fa7d-03b5-4acf-b07e-ae68b78b7c57", "address": "fa:16:3e:a2:e7:d1", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84a9fa7d-03", "ovs_interfaceid": "84a9fa7d-03b5-4acf-b07e-ae68b78b7c57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1022.238410] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520e9a45-5092-128e-fc17-7679edf8a488, 'name': SearchDatastore_Task, 'duration_secs': 0.014577} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.241045] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f840ed3-0492-402c-861a-4783383454af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.255022] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1022.255022] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b01ca1-15e9-d958-3df7-ceb8589aa878" [ 1022.255022] env[65121]: _type = "Task" [ 1022.255022] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.266756] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b01ca1-15e9-d958-3df7-ceb8589aa878, 'name': SearchDatastore_Task, 'duration_secs': 0.013886} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.267835] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.268342] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.268780] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-085ad552-ea83-4c5f-beb7-10d5c3f59133 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.281139] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1022.281139] env[65121]: value = "task-5107013" [ 1022.281139] env[65121]: _type = "Task" [ 1022.281139] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.295403] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.420751] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525513e7-00fb-36c8-aa1b-497b39867995, 'name': SearchDatastore_Task, 'duration_secs': 0.015076} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.421525] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.422280] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3/4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.422280] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3742c6f2-944a-4f08-a459-10049b787624 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.434562] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1022.434562] env[65121]: value = "task-5107014" [ 1022.434562] env[65121]: _type = "Task" [ 1022.434562] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.448610] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107014, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.456536] env[65121]: DEBUG nova.compute.manager [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1022.458007] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b53bc2-ca77-4ea0-b321-9d98226c036e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.580268] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7344ac8f-7963-45b7-ad94-eb148d3adafd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.588705] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceea0758-18e1-416e-a3d4-5decc5e0eaf3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.631892] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bc4cf9-af2f-4bae-bc02-5bc337eb9ec5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.642378] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a944a5-3147-4d92-ac66-14d33facba65 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.664495] env[65121]: DEBUG nova.compute.provider_tree [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.732309] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "refresh_cache-ec463cad-4c80-4636-bc7a-9ec298a07d96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.733366] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Instance network_info: |[{"id": "84a9fa7d-03b5-4acf-b07e-ae68b78b7c57", "address": "fa:16:3e:a2:e7:d1", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84a9fa7d-03", "ovs_interfaceid": "84a9fa7d-03b5-4acf-b07e-ae68b78b7c57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1022.733841] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:e7:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eed34ae1-5f7f-4deb-9db8-85eaa1e60c29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84a9fa7d-03b5-4acf-b07e-ae68b78b7c57', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.742893] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1022.743362] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.743802] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed41ac7c-f7cf-4a9d-be76-a956e7d0b82f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.764821] env[65121]: DEBUG nova.compute.manager [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Received event network-vif-plugged-84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1022.765187] env[65121]: DEBUG oslo_concurrency.lockutils [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Acquiring lock "ec463cad-4c80-4636-bc7a-9ec298a07d96-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1022.765550] env[65121]: DEBUG oslo_concurrency.lockutils [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1022.766100] env[65121]: DEBUG oslo_concurrency.lockutils [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.766286] env[65121]: DEBUG nova.compute.manager [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] No waiting events found dispatching network-vif-plugged-84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1022.766465] env[65121]: WARNING nova.compute.manager [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Received unexpected event network-vif-plugged-84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 for instance with vm_state building and task_state spawning. [ 1022.766756] env[65121]: DEBUG nova.compute.manager [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Received event network-changed-84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1022.766967] env[65121]: DEBUG nova.compute.manager [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Refreshing instance network info cache due to event network-changed-84a9fa7d-03b5-4acf-b07e-ae68b78b7c57. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1022.767206] env[65121]: DEBUG oslo_concurrency.lockutils [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Acquiring lock "refresh_cache-ec463cad-4c80-4636-bc7a-9ec298a07d96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.767345] env[65121]: DEBUG oslo_concurrency.lockutils [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Acquired lock "refresh_cache-ec463cad-4c80-4636-bc7a-9ec298a07d96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1022.767494] env[65121]: DEBUG nova.network.neutron [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Refreshing network info cache for port 84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1022.777137] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.777137] env[65121]: value = "task-5107016" [ 1022.777137] env[65121]: _type = "Task" [ 1022.777137] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.793573] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107016, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.797410] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107013, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.947730] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107014, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.977367] env[65121]: INFO nova.compute.manager [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] instance snapshotting [ 1022.981175] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c1a54d3-0e62-4892-95bb-8721aef93b64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.008267] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8be2413-716b-4771-8bb5-037ae9db4905 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.132588] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1023.161049] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1023.161461] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1023.161844] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1023.162191] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1023.162450] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1023.162675] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1023.162929] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.163127] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1023.163346] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1023.163728] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1023.164063] env[65121]: DEBUG nova.virt.hardware [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1023.165096] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b153c9f4-a91e-4169-8159-2a99cf382182 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.169172] env[65121]: DEBUG nova.scheduler.client.report [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1023.179961] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f68560f-ed03-4b06-be13-58e79e349710 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.272933] env[65121]: WARNING neutronclient.v2_0.client [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.274148] env[65121]: WARNING openstack [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.274806] env[65121]: WARNING openstack [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.315652] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.662473} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.320780] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.321130] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.321490] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107016, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.321711] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1758853-00f7-4c75-8d5c-d19f55476ff9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.331767] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1023.331767] env[65121]: value = "task-5107017" [ 1023.331767] env[65121]: _type = "Task" [ 1023.331767] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.341590] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.445168] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107014, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.792049} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.445485] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3/4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.445748] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.446273] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7256bb43-adca-4972-8638-699483b85ea9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.457264] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1023.457264] env[65121]: value = "task-5107019" [ 1023.457264] env[65121]: _type = "Task" [ 1023.457264] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.467232] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107019, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.517150] env[65121]: WARNING openstack [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.517150] env[65121]: WARNING openstack [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.524560] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1023.525384] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-29a979e9-94ac-4c63-9a02-5076c0ca6ae1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.536498] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1023.536498] env[65121]: value = "task-5107020" [ 1023.536498] env[65121]: _type = "Task" [ 1023.536498] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.546411] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.645851] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Successfully updated port: 6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1023.659858] env[65121]: WARNING neutronclient.v2_0.client [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.661466] env[65121]: WARNING openstack [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.661648] env[65121]: WARNING openstack [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.675771] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.654s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.675771] env[65121]: DEBUG nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1023.680263] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.893s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.680544] env[65121]: DEBUG nova.objects.instance [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'pci_requests' on Instance uuid e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.770955] env[65121]: DEBUG nova.compute.manager [req-bd8cde3b-4546-468d-96db-3c3109abe445 req-06b99301-53b1-43d1-a814-92d8f7178898 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Received event network-vif-plugged-6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1023.771108] env[65121]: DEBUG oslo_concurrency.lockutils [req-bd8cde3b-4546-468d-96db-3c3109abe445 req-06b99301-53b1-43d1-a814-92d8f7178898 service nova] Acquiring lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.771312] env[65121]: DEBUG oslo_concurrency.lockutils [req-bd8cde3b-4546-468d-96db-3c3109abe445 req-06b99301-53b1-43d1-a814-92d8f7178898 service nova] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.771474] env[65121]: DEBUG oslo_concurrency.lockutils [req-bd8cde3b-4546-468d-96db-3c3109abe445 req-06b99301-53b1-43d1-a814-92d8f7178898 service nova] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.771633] env[65121]: DEBUG nova.compute.manager [req-bd8cde3b-4546-468d-96db-3c3109abe445 req-06b99301-53b1-43d1-a814-92d8f7178898 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] No waiting events found dispatching network-vif-plugged-6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1023.771789] env[65121]: WARNING nova.compute.manager [req-bd8cde3b-4546-468d-96db-3c3109abe445 req-06b99301-53b1-43d1-a814-92d8f7178898 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Received unexpected event network-vif-plugged-6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 for instance with vm_state building and task_state spawning. [ 1023.784816] env[65121]: DEBUG nova.network.neutron [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Updated VIF entry in instance network info cache for port 84a9fa7d-03b5-4acf-b07e-ae68b78b7c57. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1023.785210] env[65121]: DEBUG nova.network.neutron [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Updating instance_info_cache with network_info: [{"id": "84a9fa7d-03b5-4acf-b07e-ae68b78b7c57", "address": "fa:16:3e:a2:e7:d1", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84a9fa7d-03", "ovs_interfaceid": "84a9fa7d-03b5-4acf-b07e-ae68b78b7c57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1023.800258] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107016, 'name': CreateVM_Task, 'duration_secs': 0.641912} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.800431] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1023.800919] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.801307] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.801477] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.801788] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1023.802317] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26bbbbd2-c0c7-401d-b63b-7aac5e34dcde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.809176] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1023.809176] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b52bb2-7f0c-9423-0872-929268aaa701" [ 1023.809176] env[65121]: _type = "Task" [ 1023.809176] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.818888] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b52bb2-7f0c-9423-0872-929268aaa701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.843201] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131113} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.843201] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.843329] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dfab35-01f9-4035-b9fa-a19e7e93d472 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.866847] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.867156] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c75d2653-8fb8-466f-86b2-f2dad5e1c47f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.888826] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1023.888826] env[65121]: value = "task-5107021" [ 1023.888826] env[65121]: _type = "Task" [ 1023.888826] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.898447] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107021, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.968496] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107019, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100325} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.968682] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.969516] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3feee9b8-137c-4595-9c75-4d1f13414431 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.992602] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3/4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.992938] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-393112bd-1e28-4289-be49-cb32ea3ef276 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.015178] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1024.015178] env[65121]: value = "task-5107022" [ 1024.015178] env[65121]: _type = "Task" [ 1024.015178] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.025302] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107022, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.049042] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.149649] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "refresh_cache-abd87ca7-5d95-4b22-8d2a-1f37abd18f86" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.149944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "refresh_cache-abd87ca7-5d95-4b22-8d2a-1f37abd18f86" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1024.150204] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1024.183768] env[65121]: DEBUG nova.compute.utils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1024.187487] env[65121]: DEBUG nova.objects.instance [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'numa_topology' on Instance uuid e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.189520] env[65121]: DEBUG nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1024.189764] env[65121]: DEBUG nova.network.neutron [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1024.190293] env[65121]: WARNING neutronclient.v2_0.client [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.190669] env[65121]: WARNING neutronclient.v2_0.client [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.191427] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.191883] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.238969] env[65121]: DEBUG oslo_concurrency.lockutils [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1024.239571] env[65121]: DEBUG oslo_concurrency.lockutils [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.239984] env[65121]: DEBUG nova.objects.instance [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'flavor' on Instance uuid 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.249981] env[65121]: DEBUG nova.policy [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8931a266fddc4346b0c5243608a39bff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00383aa3355e438cb703c2b86c7917f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1024.288933] env[65121]: DEBUG oslo_concurrency.lockutils [req-7d1db3d4-07fe-4c64-88ff-f1225921db21 req-efda6a7d-b103-4975-b8d4-f0f60afa0354 service nova] Releasing lock "refresh_cache-ec463cad-4c80-4636-bc7a-9ec298a07d96" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.322748] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b52bb2-7f0c-9423-0872-929268aaa701, 'name': SearchDatastore_Task, 'duration_secs': 0.015522} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.323326] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.324110] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.324290] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.324426] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1024.324659] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.325000] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c070ba1-0ed9-4bdb-bd57-750a2ca6a04b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.336160] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.336402] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1024.337381] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53d1974a-c6ea-4507-87ad-9b106c6f177c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.344376] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1024.344376] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5235012f-001b-145b-6f1a-11bfbca1802c" [ 1024.344376] env[65121]: _type = "Task" [ 1024.344376] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.355933] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5235012f-001b-145b-6f1a-11bfbca1802c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.400844] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107021, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.525732] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107022, 'name': ReconfigVM_Task, 'duration_secs': 0.307617} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.526021] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3/4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.526714] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a40cadc7-ddd1-494c-95fe-5297749b7b78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.534547] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1024.534547] env[65121]: value = "task-5107023" [ 1024.534547] env[65121]: _type = "Task" [ 1024.534547] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.546765] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107023, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.550173] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107020, 'name': CreateSnapshot_Task, 'duration_secs': 0.813719} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.550440] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1024.551218] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2019e5-0170-47c7-a32e-7b8e19a17970 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.629929] env[65121]: DEBUG nova.network.neutron [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Successfully created port: 53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1024.653017] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.653427] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.695526] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1024.700642] env[65121]: DEBUG nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1024.704413] env[65121]: INFO nova.compute.claims [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.736751] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.736996] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.745862] env[65121]: WARNING neutronclient.v2_0.client [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.746661] env[65121]: WARNING openstack [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.747071] env[65121]: WARNING openstack [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.751989] env[65121]: DEBUG nova.objects.instance [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'pci_requests' on Instance uuid 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.823200] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.823200] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.823200] env[65121]: WARNING openstack [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.856381] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5235012f-001b-145b-6f1a-11bfbca1802c, 'name': SearchDatastore_Task, 'duration_secs': 0.015985} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.857251] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70db3b67-d3e0-4b1b-864a-6acaa791e22d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.863228] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1024.863228] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52acf31a-5593-25fd-c32c-67302876f49a" [ 1024.863228] env[65121]: _type = "Task" [ 1024.863228] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.872918] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52acf31a-5593-25fd-c32c-67302876f49a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.894439] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1024.894668] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993517', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'name': 'volume-7bcf71f4-0c34-441a-9533-931791d14ff6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59ceb0fd-1fb5-4c90-963d-fe76b9740d29', 'attached_at': '', 'detached_at': '', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'serial': '7bcf71f4-0c34-441a-9533-931791d14ff6'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1024.895833] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929f98f8-8d53-40bc-a554-92f8ed22404c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.918467] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1cc192-e232-4846-bd92-91fdba65d510 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.921699] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107021, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.945943] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] volume-7bcf71f4-0c34-441a-9533-931791d14ff6/volume-7bcf71f4-0c34-441a-9533-931791d14ff6.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.948889] env[65121]: DEBUG nova.network.neutron [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Updating instance_info_cache with network_info: [{"id": "6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35", "address": "fa:16:3e:34:f3:e6", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d471ad5-5d", "ovs_interfaceid": "6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1024.950311] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-999e66e4-a25f-411f-8af6-6a38422d1c17 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.966285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "refresh_cache-abd87ca7-5d95-4b22-8d2a-1f37abd18f86" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1024.966634] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Instance network_info: |[{"id": "6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35", "address": "fa:16:3e:34:f3:e6", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d471ad5-5d", "ovs_interfaceid": "6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1024.967326] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:f3:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eed34ae1-5f7f-4deb-9db8-85eaa1e60c29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.974700] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1024.975649] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1024.975892] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f62482e2-0ba8-470f-9e15-13530d1b2ba8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.992824] env[65121]: DEBUG oslo_vmware.api [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1024.992824] env[65121]: value = "task-5107024" [ 1024.992824] env[65121]: _type = "Task" [ 1024.992824] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.003049] env[65121]: DEBUG oslo_vmware.api [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107024, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.004492] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.004492] env[65121]: value = "task-5107025" [ 1025.004492] env[65121]: _type = "Task" [ 1025.004492] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.014961] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107025, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.045320] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107023, 'name': Rename_Task, 'duration_secs': 0.148207} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.045801] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.046037] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5241804-deb4-4f5e-94fa-f639977c98d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.053371] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1025.053371] env[65121]: value = "task-5107026" [ 1025.053371] env[65121]: _type = "Task" [ 1025.053371] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.062072] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107026, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.074490] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1025.074832] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-00febb50-58bd-4c95-887f-9f386ca3bc0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.084668] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1025.084668] env[65121]: value = "task-5107027" [ 1025.084668] env[65121]: _type = "Task" [ 1025.084668] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.094610] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107027, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.256549] env[65121]: DEBUG nova.objects.base [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Object Instance<23d6f632-8918-46af-b239-08a9615dfbec> lazy-loaded attributes: flavor,pci_requests {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1025.257437] env[65121]: DEBUG nova.network.neutron [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1025.258347] env[65121]: WARNING neutronclient.v2_0.client [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.258912] env[65121]: WARNING neutronclient.v2_0.client [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.260117] env[65121]: WARNING openstack [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1025.261147] env[65121]: WARNING openstack [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1025.373437] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52acf31a-5593-25fd-c32c-67302876f49a, 'name': SearchDatastore_Task, 'duration_secs': 0.014415} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.373807] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1025.374214] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] ec463cad-4c80-4636-bc7a-9ec298a07d96/ec463cad-4c80-4636-bc7a-9ec298a07d96.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1025.374564] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0517a335-53c0-458e-b952-f3d0a86d687d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.382117] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1025.382117] env[65121]: value = "task-5107028" [ 1025.382117] env[65121]: _type = "Task" [ 1025.382117] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.392059] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107028, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.402962] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107021, 'name': ReconfigVM_Task, 'duration_secs': 1.324691} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.403141] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Reconfigured VM instance instance-00000057 to attach disk [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf/bad70910-58a1-4e35-8d0c-3bd7c3a30abf.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.403882] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df308e51-1437-4c8d-9bb5-0fc9cae85cc3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.413566] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1025.413566] env[65121]: value = "task-5107029" [ 1025.413566] env[65121]: _type = "Task" [ 1025.413566] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.428162] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107029, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.439129] env[65121]: DEBUG oslo_concurrency.lockutils [None req-41e53656-4d9c-4b2a-89e9-fd8de83582a2 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.200s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.506535] env[65121]: DEBUG oslo_vmware.api [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107024, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.516732] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107025, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.568561] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107026, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.598018] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107027, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.721920] env[65121]: DEBUG nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1025.753361] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1025.753361] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1025.753635] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1025.753635] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1025.753704] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1025.753834] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1025.754071] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.754287] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1025.754477] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1025.754670] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1025.754866] env[65121]: DEBUG nova.virt.hardware [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1025.755910] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792a877f-6ab2-4fa3-a723-b975be0f6099 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.766942] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31a7e02-36fe-4181-9499-eae7e8b852dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.787383] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1025.787626] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993520', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'name': 'volume-79bb9f39-f959-4b03-8058-0e7401b28853', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6fe86deb-0903-4769-a05c-b7d0acec0103', 'attached_at': '', 'detached_at': '', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'serial': '79bb9f39-f959-4b03-8058-0e7401b28853'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1025.788564] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5914b75-966d-47de-9b6a-35c033002cf2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.813448] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9920ad6b-5e99-4c4e-a784-aa9268c64739 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.843603] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] volume-79bb9f39-f959-4b03-8058-0e7401b28853/volume-79bb9f39-f959-4b03-8058-0e7401b28853.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.847000] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a990ae0b-788d-436e-a65c-71af4d5d7003 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.870333] env[65121]: DEBUG oslo_vmware.api [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1025.870333] env[65121]: value = "task-5107030" [ 1025.870333] env[65121]: _type = "Task" [ 1025.870333] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.884994] env[65121]: DEBUG oslo_vmware.api [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107030, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.899650] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107028, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.926108] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107029, 'name': Rename_Task, 'duration_secs': 0.210629} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.926504] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.926770] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd8555dd-1e76-42cf-8354-bae90e34fe0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.939477] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1025.939477] env[65121]: value = "task-5107031" [ 1025.939477] env[65121]: _type = "Task" [ 1025.939477] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.960576] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107031, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.007734] env[65121]: DEBUG oslo_vmware.api [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107024, 'name': ReconfigVM_Task, 'duration_secs': 0.876037} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.013690] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Reconfigured VM instance instance-00000050 to attach disk [datastore1] volume-7bcf71f4-0c34-441a-9533-931791d14ff6/volume-7bcf71f4-0c34-441a-9533-931791d14ff6.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.019727] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a69686ac-13c3-446d-8b4a-ab87e6bf6d87 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.040942] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107025, 'name': CreateVM_Task, 'duration_secs': 0.547461} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.041970] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1026.042335] env[65121]: DEBUG oslo_vmware.api [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1026.042335] env[65121]: value = "task-5107032" [ 1026.042335] env[65121]: _type = "Task" [ 1026.042335] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.043257] env[65121]: WARNING neutronclient.v2_0.client [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1026.043257] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.043524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.043649] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1026.043959] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6271848e-a6b0-4102-9ead-e3646bedc207 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.063855] env[65121]: DEBUG oslo_vmware.api [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107032, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.064292] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1026.064292] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c67a30-93a8-2fe9-631e-9d002a074fb5" [ 1026.064292] env[65121]: _type = "Task" [ 1026.064292] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.076387] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107026, 'name': PowerOnVM_Task, 'duration_secs': 0.5557} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.077175] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.077547] env[65121]: INFO nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Took 8.18 seconds to spawn the instance on the hypervisor. [ 1026.077651] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1026.078568] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8eeb830-96e6-4c73-9630-380dc8c72939 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.087976] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c67a30-93a8-2fe9-631e-9d002a074fb5, 'name': SearchDatastore_Task, 'duration_secs': 0.017936} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.089040] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1026.089279] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.089520] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.089658] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.089830] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.094923] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a128e3ad-be85-4240-a922-cd4dc6d02973 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.108866] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107027, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.111797] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.111899] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1026.112641] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3ef4a30-d553-4427-b9e0-697bcdef077e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.120944] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1026.120944] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5270b687-292e-1a7e-9fb4-eb62da44c23c" [ 1026.120944] env[65121]: _type = "Task" [ 1026.120944] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.131501] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5270b687-292e-1a7e-9fb4-eb62da44c23c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.228516] env[65121]: DEBUG nova.network.neutron [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Successfully updated port: 53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1026.261226] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147feb9b-67fe-40a2-b796-a6febb072e63 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.271194] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda70ebb-3d2c-431b-9a44-1fe1a64787c5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.278032] env[65121]: DEBUG nova.compute.manager [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Received event network-changed-6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1026.278426] env[65121]: DEBUG nova.compute.manager [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Refreshing instance network info cache due to event network-changed-6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1026.278682] env[65121]: DEBUG oslo_concurrency.lockutils [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Acquiring lock "refresh_cache-abd87ca7-5d95-4b22-8d2a-1f37abd18f86" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.278887] env[65121]: DEBUG oslo_concurrency.lockutils [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Acquired lock "refresh_cache-abd87ca7-5d95-4b22-8d2a-1f37abd18f86" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.279177] env[65121]: DEBUG nova.network.neutron [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Refreshing network info cache for port 6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1026.319805] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45490e2-970c-4f70-882f-93a1245f13a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.332051] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d5b01c-dcf6-4ba5-b229-3f2784edab9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.348116] env[65121]: DEBUG nova.compute.provider_tree [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.383033] env[65121]: DEBUG oslo_vmware.api [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.395213] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107028, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613597} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.395506] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] ec463cad-4c80-4636-bc7a-9ec298a07d96/ec463cad-4c80-4636-bc7a-9ec298a07d96.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1026.395794] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.396095] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6dc39f75-d725-43e6-8a34-cb26c4d849f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.405670] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1026.405670] env[65121]: value = "task-5107033" [ 1026.405670] env[65121]: _type = "Task" [ 1026.405670] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.417707] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.452157] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107031, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.555229] env[65121]: DEBUG oslo_vmware.api [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107032, 'name': ReconfigVM_Task, 'duration_secs': 0.202953} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.555650] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993517', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'name': 'volume-7bcf71f4-0c34-441a-9533-931791d14ff6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59ceb0fd-1fb5-4c90-963d-fe76b9740d29', 'attached_at': '', 'detached_at': '', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'serial': '7bcf71f4-0c34-441a-9533-931791d14ff6'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1026.598882] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107027, 'name': CloneVM_Task, 'duration_secs': 1.480639} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.599265] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Created linked-clone VM from snapshot [ 1026.600131] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8db5917-fccf-4cf1-ba11-572d4d400d07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.609696] env[65121]: INFO nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Took 44.87 seconds to build instance. [ 1026.615674] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Uploading image 4b58f9f6-64be-4173-9dc8-2f49a8476dc8 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1026.631451] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5270b687-292e-1a7e-9fb4-eb62da44c23c, 'name': SearchDatastore_Task, 'duration_secs': 0.014804} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.633055] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1026.634193] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-81f035a7-c8c0-4efe-8f41-2d367c2c12ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.638039] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33a33c54-0927-4663-8bf3-33c2f6681b0a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.645150] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1026.645150] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52789fa8-8fa6-45cd-5370-a075b972b22c" [ 1026.645150] env[65121]: _type = "Task" [ 1026.645150] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.649829] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1026.649829] env[65121]: value = "task-5107034" [ 1026.649829] env[65121]: _type = "Task" [ 1026.649829] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.657599] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52789fa8-8fa6-45cd-5370-a075b972b22c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.663067] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107034, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.731516] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.731748] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.731961] env[65121]: DEBUG nova.network.neutron [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1026.782857] env[65121]: WARNING neutronclient.v2_0.client [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1026.783576] env[65121]: WARNING openstack [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.783921] env[65121]: WARNING openstack [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.852274] env[65121]: DEBUG nova.scheduler.client.report [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1026.883461] env[65121]: DEBUG oslo_vmware.api [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107030, 'name': ReconfigVM_Task, 'duration_secs': 0.674799} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.883776] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Reconfigured VM instance instance-0000004f to attach disk [datastore1] volume-79bb9f39-f959-4b03-8058-0e7401b28853/volume-79bb9f39-f959-4b03-8058-0e7401b28853.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.893701] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed7f12ca-137e-48aa-b5a7-d896c509ecc3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.912168] env[65121]: DEBUG oslo_vmware.api [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1026.912168] env[65121]: value = "task-5107035" [ 1026.912168] env[65121]: _type = "Task" [ 1026.912168] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.919706] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080451} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.920422] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.921249] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98912bff-a6f8-43a5-bbd5-1ff5ffe200b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.928308] env[65121]: WARNING openstack [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.928737] env[65121]: WARNING openstack [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.935877] env[65121]: DEBUG oslo_vmware.api [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107035, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.961067] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] ec463cad-4c80-4636-bc7a-9ec298a07d96/ec463cad-4c80-4636-bc7a-9ec298a07d96.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.963953] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-063b0c28-542c-4569-8066-16afa138ee5a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.991782] env[65121]: DEBUG oslo_vmware.api [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107031, 'name': PowerOnVM_Task, 'duration_secs': 0.684169} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.993321] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.993541] env[65121]: DEBUG nova.compute.manager [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1026.993880] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1026.993880] env[65121]: value = "task-5107036" [ 1026.993880] env[65121]: _type = "Task" [ 1026.993880] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.994669] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3d7abd-aaa1-4f1d-91b4-e5acc398a126 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.012357] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107036, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.030781] env[65121]: WARNING neutronclient.v2_0.client [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.031630] env[65121]: WARNING openstack [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.032075] env[65121]: WARNING openstack [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.120783] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.392s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.130103] env[65121]: DEBUG nova.network.neutron [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Updated VIF entry in instance network info cache for port 6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1027.130239] env[65121]: DEBUG nova.network.neutron [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Updating instance_info_cache with network_info: [{"id": "6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35", "address": "fa:16:3e:34:f3:e6", "network": {"id": "c8032a02-5737-479d-a631-61b62d113198", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1958685449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4feb04f47a97427c8a2bc97883332e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eed34ae1-5f7f-4deb-9db8-85eaa1e60c29", "external-id": "nsx-vlan-transportzone-780", "segmentation_id": 780, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d471ad5-5d", "ovs_interfaceid": "6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1027.166574] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107034, 'name': Destroy_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.166856] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52789fa8-8fa6-45cd-5370-a075b972b22c, 'name': SearchDatastore_Task, 'duration_secs': 0.021366} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.167115] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.167420] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] abd87ca7-5d95-4b22-8d2a-1f37abd18f86/abd87ca7-5d95-4b22-8d2a-1f37abd18f86.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1027.167653] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f40242d-ce15-4d81-b2e3-67e6e2563988 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.176745] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1027.176745] env[65121]: value = "task-5107037" [ 1027.176745] env[65121]: _type = "Task" [ 1027.176745] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.187334] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.235260] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.235681] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.288641] env[65121]: DEBUG nova.network.neutron [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1027.312699] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.313223] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.358914] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.679s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.358914] env[65121]: WARNING neutronclient.v2_0.client [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.361225] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.611s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.361421] env[65121]: DEBUG nova.objects.instance [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lazy-loading 'resources' on Instance uuid e46c88ed-2c2e-432d-987d-0e4c99c28e42 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.371169] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.371433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.371711] env[65121]: DEBUG nova.objects.instance [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'flavor' on Instance uuid 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.391774] env[65121]: WARNING neutronclient.v2_0.client [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.392486] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.392840] env[65121]: WARNING openstack [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.406905] env[65121]: INFO nova.network.neutron [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating port 71e0942f-5026-4128-ba81-16311feb9b3e with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1027.423872] env[65121]: DEBUG oslo_vmware.api [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107035, 'name': ReconfigVM_Task, 'duration_secs': 0.214444} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.424247] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993520', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'name': 'volume-79bb9f39-f959-4b03-8058-0e7401b28853', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6fe86deb-0903-4769-a05c-b7d0acec0103', 'attached_at': '', 'detached_at': '', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'serial': '79bb9f39-f959-4b03-8058-0e7401b28853'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1027.509779] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107036, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.521735] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.537667] env[65121]: DEBUG nova.network.neutron [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updating instance_info_cache with network_info: [{"id": "53616727-ab6a-4f06-a597-0fc068a18955", "address": "fa:16:3e:90:9a:47", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53616727-ab", "ovs_interfaceid": "53616727-ab6a-4f06-a597-0fc068a18955", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1027.598277] env[65121]: DEBUG nova.objects.instance [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'flavor' on Instance uuid 59ceb0fd-1fb5-4c90-963d-fe76b9740d29 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.633800] env[65121]: DEBUG oslo_concurrency.lockutils [req-7c7350de-f296-4410-ab87-4e52cf772843 req-54f74574-012b-491b-bd75-dbad00388400 service nova] Releasing lock "refresh_cache-abd87ca7-5d95-4b22-8d2a-1f37abd18f86" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.665570] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107034, 'name': Destroy_Task, 'duration_secs': 0.703044} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.665929] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Destroyed the VM [ 1027.666281] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1027.666554] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a459902e-5d17-4493-9274-75c9df879bad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.674429] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1027.674429] env[65121]: value = "task-5107038" [ 1027.674429] env[65121]: _type = "Task" [ 1027.674429] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.687318] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107038, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.690715] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107037, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.876047] env[65121]: WARNING neutronclient.v2_0.client [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.876705] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.877071] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.981528] env[65121]: DEBUG nova.objects.instance [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'pci_requests' on Instance uuid 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.011676] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107036, 'name': ReconfigVM_Task, 'duration_secs': 0.765666} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.011961] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Reconfigured VM instance instance-0000005a to attach disk [datastore1] ec463cad-4c80-4636-bc7a-9ec298a07d96/ec463cad-4c80-4636-bc7a-9ec298a07d96.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.012818] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ad676bb-8ac4-4f07-b55c-b9806bbd070a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.023210] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1028.023210] env[65121]: value = "task-5107039" [ 1028.023210] env[65121]: _type = "Task" [ 1028.023210] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.037084] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107039, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.042206] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.042785] env[65121]: DEBUG nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Instance network_info: |[{"id": "53616727-ab6a-4f06-a597-0fc068a18955", "address": "fa:16:3e:90:9a:47", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53616727-ab", "ovs_interfaceid": "53616727-ab6a-4f06-a597-0fc068a18955", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1028.043936] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:9a:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53616727-ab6a-4f06-a597-0fc068a18955', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.052987] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1028.053697] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.056963] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6086fa4e-9ae7-41b7-b6df-1ab3f1602a46 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.084108] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.084108] env[65121]: value = "task-5107040" [ 1028.084108] env[65121]: _type = "Task" [ 1028.084108] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.095706] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107040, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.103869] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6165d939-da82-4c57-943a-0c4f10ebb1ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.858s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.195111] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.75068} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.199216] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] abd87ca7-5d95-4b22-8d2a-1f37abd18f86/abd87ca7-5d95-4b22-8d2a-1f37abd18f86.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1028.199503] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1028.199806] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107038, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.203060] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f45ca7b-bbf0-447d-9140-249aeb9aceac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.213609] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1028.213609] env[65121]: value = "task-5107041" [ 1028.213609] env[65121]: _type = "Task" [ 1028.213609] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.229839] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107041, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.373230] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55715123-a367-4422-b868-8574665601c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.382731] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a487109-4d8b-4632-bf66-875c4775bf14 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.428452] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc037dc-40d2-449a-9591-7d7947f84779 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.437576] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa32b3c-9c3e-4e88-bf58-0a902455568a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.454801] env[65121]: DEBUG nova.compute.provider_tree [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.476828] env[65121]: DEBUG nova.objects.instance [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.483885] env[65121]: DEBUG nova.objects.base [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Object Instance<23d6f632-8918-46af-b239-08a9615dfbec> lazy-loaded attributes: flavor,pci_requests {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1028.483885] env[65121]: DEBUG nova.network.neutron [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1028.484211] env[65121]: WARNING neutronclient.v2_0.client [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.484557] env[65121]: WARNING neutronclient.v2_0.client [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.485218] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.485579] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.535293] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107039, 'name': Rename_Task, 'duration_secs': 0.24887} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.538881] env[65121]: DEBUG nova.policy [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1028.540774] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1028.541385] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41149a7e-55c4-48eb-9bb3-9fffb8da5d3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.549071] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1028.549071] env[65121]: value = "task-5107042" [ 1028.549071] env[65121]: _type = "Task" [ 1028.549071] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.558074] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.595682] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107040, 'name': CreateVM_Task, 'duration_secs': 0.43967} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.595901] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.596529] env[65121]: WARNING neutronclient.v2_0.client [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.596930] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.597116] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.597470] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1028.597723] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e332936-b771-41cc-9b0d-162da012668f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.604346] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1028.604346] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d38ea2-5de3-350b-8cb9-6c03d848ba96" [ 1028.604346] env[65121]: _type = "Task" [ 1028.604346] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.615218] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d38ea2-5de3-350b-8cb9-6c03d848ba96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.686081] env[65121]: DEBUG oslo_vmware.api [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107038, 'name': RemoveSnapshot_Task, 'duration_secs': 0.75371} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.686081] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1028.724116] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107041, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097959} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.724459] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.725409] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b2c907-1c40-4701-a37e-ffe6ed07f443 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.755138] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] abd87ca7-5d95-4b22-8d2a-1f37abd18f86/abd87ca7-5d95-4b22-8d2a-1f37abd18f86.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.756617] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-158ca2fc-c54e-432e-896a-e5a70b330c91 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.789703] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1028.789703] env[65121]: value = "task-5107043" [ 1028.789703] env[65121]: _type = "Task" [ 1028.789703] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.800130] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107043, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.899315] env[65121]: DEBUG nova.network.neutron [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Successfully created port: 15f422aa-08f5-4473-b1ac-92eb03870e58 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1028.959212] env[65121]: DEBUG nova.scheduler.client.report [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1028.981789] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a160951-200a-44b8-a768-210aafb2b8dc tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.327s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.005953] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.006096] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.006343] env[65121]: DEBUG nova.network.neutron [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1029.061830] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107042, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.115913] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d38ea2-5de3-350b-8cb9-6c03d848ba96, 'name': SearchDatastore_Task, 'duration_secs': 0.015801} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.116275] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.116495] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.116723] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.116862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.117043] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.117322] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-975d4420-e8ff-4ff7-a515-4fb0df998cfa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.132022] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.132022] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.132022] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca224924-8984-420f-b9fc-2f4e4aad2685 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.139460] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1029.139460] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52481385-7e33-e0f3-f912-04a69045f5ef" [ 1029.139460] env[65121]: _type = "Task" [ 1029.139460] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.151685] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52481385-7e33-e0f3-f912-04a69045f5ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.192587] env[65121]: WARNING nova.compute.manager [None req-c541b151-8aa2-484f-92af-df5303f52a15 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Image not found during snapshot: nova.exception.ImageNotFound: Image 4b58f9f6-64be-4173-9dc8-2f49a8476dc8 could not be found. [ 1029.302403] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107043, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.464943] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.104s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.468496] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.136s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.471539] env[65121]: INFO nova.compute.claims [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1029.500154] env[65121]: INFO nova.scheduler.client.report [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleted allocations for instance e46c88ed-2c2e-432d-987d-0e4c99c28e42 [ 1029.510734] env[65121]: WARNING neutronclient.v2_0.client [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.512806] env[65121]: WARNING openstack [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.512806] env[65121]: WARNING openstack [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.565270] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107042, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.652751] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52481385-7e33-e0f3-f912-04a69045f5ef, 'name': SearchDatastore_Task, 'duration_secs': 0.018949} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.653837] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66422bb9-a314-4642-8c1e-305aaef167e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.661960] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1029.661960] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cdc5ba-d98a-30c3-5d6a-3b55d074f416" [ 1029.661960] env[65121]: _type = "Task" [ 1029.661960] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.671842] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cdc5ba-d98a-30c3-5d6a-3b55d074f416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.700427] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.700427] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.700427] env[65121]: DEBUG nova.compute.manager [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1029.700427] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394e8dd0-99b4-4e3c-9e8e-ce15d154c291 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.710052] env[65121]: DEBUG nova.compute.manager [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1029.710809] env[65121]: DEBUG nova.objects.instance [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.792830] env[65121]: DEBUG nova.compute.manager [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Received event network-vif-plugged-53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1029.793046] env[65121]: DEBUG oslo_concurrency.lockutils [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Acquiring lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.793252] env[65121]: DEBUG oslo_concurrency.lockutils [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.793406] env[65121]: DEBUG oslo_concurrency.lockutils [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.793557] env[65121]: DEBUG nova.compute.manager [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] No waiting events found dispatching network-vif-plugged-53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1029.794750] env[65121]: WARNING nova.compute.manager [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Received unexpected event network-vif-plugged-53616727-ab6a-4f06-a597-0fc068a18955 for instance with vm_state building and task_state spawning. [ 1029.794750] env[65121]: DEBUG nova.compute.manager [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Received event network-changed-53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1029.794750] env[65121]: DEBUG nova.compute.manager [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Refreshing instance network info cache due to event network-changed-53616727-ab6a-4f06-a597-0fc068a18955. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1029.794750] env[65121]: DEBUG oslo_concurrency.lockutils [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Acquiring lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.794750] env[65121]: DEBUG oslo_concurrency.lockutils [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Acquired lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.794750] env[65121]: DEBUG nova.network.neutron [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Refreshing network info cache for port 53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1029.809399] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107043, 'name': ReconfigVM_Task, 'duration_secs': 0.766052} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.810684] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Reconfigured VM instance instance-0000005b to attach disk [datastore2] abd87ca7-5d95-4b22-8d2a-1f37abd18f86/abd87ca7-5d95-4b22-8d2a-1f37abd18f86.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.811755] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d75363d4-7168-4601-9353-7fba6e63ac5d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.821286] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1029.821286] env[65121]: value = "task-5107044" [ 1029.821286] env[65121]: _type = "Task" [ 1029.821286] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.831306] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107044, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.928887] env[65121]: WARNING openstack [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.929382] env[65121]: WARNING openstack [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.008704] env[65121]: DEBUG oslo_concurrency.lockutils [None req-74ea5d37-ca0c-4997-a86e-cad483b10e79 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "e46c88ed-2c2e-432d-987d-0e4c99c28e42" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.688s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.013042] env[65121]: WARNING neutronclient.v2_0.client [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.013042] env[65121]: WARNING openstack [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.013042] env[65121]: WARNING openstack [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.060835] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107042, 'name': PowerOnVM_Task, 'duration_secs': 1.018568} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.061167] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.061371] env[65121]: INFO nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Took 9.56 seconds to spawn the instance on the hypervisor. [ 1030.061550] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1030.062542] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52575be2-442d-4728-ab60-ab9600aac5f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.139203] env[65121]: DEBUG nova.network.neutron [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e0942f-50", "ovs_interfaceid": "71e0942f-5026-4128-ba81-16311feb9b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.176953] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cdc5ba-d98a-30c3-5d6a-3b55d074f416, 'name': SearchDatastore_Task, 'duration_secs': 0.018456} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.177277] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.177713] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.177906] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0f84874-4f26-4dbd-b734-2a114d1114ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.185668] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1030.185668] env[65121]: value = "task-5107045" [ 1030.185668] env[65121]: _type = "Task" [ 1030.185668] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.195123] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.302083] env[65121]: WARNING neutronclient.v2_0.client [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.302631] env[65121]: WARNING openstack [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.303015] env[65121]: WARNING openstack [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.333864] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107044, 'name': Rename_Task, 'duration_secs': 0.306736} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.334372] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1030.334826] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f537b567-3435-42e2-9096-5570ae107c0d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.344707] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1030.344707] env[65121]: value = "task-5107046" [ 1030.344707] env[65121]: _type = "Task" [ 1030.344707] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.355654] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107046, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.488792] env[65121]: WARNING openstack [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.488792] env[65121]: WARNING openstack [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.570292] env[65121]: WARNING neutronclient.v2_0.client [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.571456] env[65121]: WARNING openstack [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.574081] env[65121]: WARNING openstack [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.589019] env[65121]: DEBUG nova.compute.manager [req-50c00478-2d5b-4a00-8b41-a8adcb90631c req-fa9e4e07-5321-48be-b15e-f8b4cd944a93 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-vif-plugged-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1030.589019] env[65121]: DEBUG oslo_concurrency.lockutils [req-50c00478-2d5b-4a00-8b41-a8adcb90631c req-fa9e4e07-5321-48be-b15e-f8b4cd944a93 service nova] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.589019] env[65121]: DEBUG oslo_concurrency.lockutils [req-50c00478-2d5b-4a00-8b41-a8adcb90631c req-fa9e4e07-5321-48be-b15e-f8b4cd944a93 service nova] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.589728] env[65121]: DEBUG oslo_concurrency.lockutils [req-50c00478-2d5b-4a00-8b41-a8adcb90631c req-fa9e4e07-5321-48be-b15e-f8b4cd944a93 service nova] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.590215] env[65121]: DEBUG nova.compute.manager [req-50c00478-2d5b-4a00-8b41-a8adcb90631c req-fa9e4e07-5321-48be-b15e-f8b4cd944a93 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] No waiting events found dispatching network-vif-plugged-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1030.590580] env[65121]: WARNING nova.compute.manager [req-50c00478-2d5b-4a00-8b41-a8adcb90631c req-fa9e4e07-5321-48be-b15e-f8b4cd944a93 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received unexpected event network-vif-plugged-71e0942f-5026-4128-ba81-16311feb9b3e for instance with vm_state shelved_offloaded and task_state spawning. [ 1030.611380] env[65121]: DEBUG nova.network.neutron [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Successfully updated port: 15f422aa-08f5-4473-b1ac-92eb03870e58 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1030.617831] env[65121]: INFO nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Took 48.83 seconds to build instance. [ 1030.643281] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.686012] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='496cec69643f7ad73d94b1ea3ccd7df4',container_format='bare',created_at=2025-12-12T14:26:01Z,direct_url=,disk_format='vmdk',id=91345b02-0e33-4bb1-bc26-4933119dd392,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-573163293-shelved',owner='ad1aea30d62c45e193c5a54c429ce7ba',properties=ImageMetaProps,protected=,size=31673344,status='active',tags=,updated_at=2025-12-12T14:26:17Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1030.686339] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1030.686506] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1030.686715] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1030.686878] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1030.687010] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1030.687250] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.687405] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1030.687537] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1030.687684] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1030.687853] env[65121]: DEBUG nova.virt.hardware [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1030.688863] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f55a0e3-ae14-4c85-ad78-bc88848a5d60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.698548] env[65121]: DEBUG nova.network.neutron [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updated VIF entry in instance network info cache for port 53616727-ab6a-4f06-a597-0fc068a18955. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1030.698928] env[65121]: DEBUG nova.network.neutron [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updating instance_info_cache with network_info: [{"id": "53616727-ab6a-4f06-a597-0fc068a18955", "address": "fa:16:3e:90:9a:47", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53616727-ab", "ovs_interfaceid": "53616727-ab6a-4f06-a597-0fc068a18955", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.712312] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b31f98-f527-451e-9a28-41d276f373a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.717160] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107045, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.721035] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.722461] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08743842-4387-42de-b144-74348fd59b61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.737172] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:a9:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a64108f9-df0a-4feb-bbb5-97f5841c356c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71e0942f-5026-4128-ba81-16311feb9b3e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.745702] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1030.751422] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.751831] env[65121]: DEBUG oslo_vmware.api [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1030.751831] env[65121]: value = "task-5107047" [ 1030.751831] env[65121]: _type = "Task" [ 1030.751831] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.752550] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.752794] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.753221] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.753388] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.753600] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.755361] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d04edee-6a39-446b-bcd6-86f0d8804d6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.773995] env[65121]: INFO nova.compute.manager [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Terminating instance [ 1030.793240] env[65121]: DEBUG oslo_vmware.api [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107047, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.795921] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.795921] env[65121]: value = "task-5107048" [ 1030.795921] env[65121]: _type = "Task" [ 1030.795921] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.808944] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107048, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.857598] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107046, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.001086] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdc7437-d523-4d63-83b4-bc8e14cdb06c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.013773] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33801f4-335d-4e79-9cab-af429603ecef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.048604] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ce5232-9311-4bf8-9155-75490aab6670 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.057795] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620cb4f9-4e4c-430c-92b7-e1159327de9c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.076658] env[65121]: DEBUG nova.compute.provider_tree [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.117045] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.117205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.117364] env[65121]: DEBUG nova.network.neutron [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1031.120785] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.345s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.199467] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.624184} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.199776] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.199968] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.200246] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62e552b7-4744-45b5-9cbd-a57eb1703b6d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.204782] env[65121]: DEBUG oslo_concurrency.lockutils [req-cd015390-e9a5-4c95-8a8e-c6bf427ffdee req-a0ccc56f-ac73-468a-9281-b6a01b1b8599 service nova] Releasing lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.208758] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1031.208758] env[65121]: value = "task-5107049" [ 1031.208758] env[65121]: _type = "Task" [ 1031.208758] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.219048] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107049, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.288253] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "refresh_cache-bad70910-58a1-4e35-8d0c-3bd7c3a30abf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.288643] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquired lock "refresh_cache-bad70910-58a1-4e35-8d0c-3bd7c3a30abf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.288643] env[65121]: DEBUG nova.network.neutron [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1031.290388] env[65121]: DEBUG oslo_vmware.api [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107047, 'name': PowerOffVM_Task, 'duration_secs': 0.35067} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.290544] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.290684] env[65121]: DEBUG nova.compute.manager [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1031.292074] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f3215d-b3d9-4540-8c7d-62d572d488cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.309975] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107048, 'name': CreateVM_Task, 'duration_secs': 0.502985} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.313803] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.315435] env[65121]: WARNING neutronclient.v2_0.client [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.316642] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.316642] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.316642] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1031.317031] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b924ed05-a0b5-41c7-850d-5170b7ded7c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.325653] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1031.325653] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524a960a-fc5c-a582-dedd-6470554f21cb" [ 1031.325653] env[65121]: _type = "Task" [ 1031.325653] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.340259] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524a960a-fc5c-a582-dedd-6470554f21cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.358039] env[65121]: DEBUG oslo_vmware.api [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107046, 'name': PowerOnVM_Task, 'duration_secs': 1.008871} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.359285] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1031.360147] env[65121]: INFO nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Took 8.23 seconds to spawn the instance on the hypervisor. [ 1031.360147] env[65121]: DEBUG nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1031.360825] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabd04cd-e2e6-4640-80fb-4ccc56067f9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.580174] env[65121]: DEBUG nova.scheduler.client.report [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1031.620778] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.620778] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.651048] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271e88b-434f-414f-2a98-2cff23d7bfe9/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1031.652311] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad56d09-ea23-4870-9dae-a009ace419a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.658999] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271e88b-434f-414f-2a98-2cff23d7bfe9/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1031.659178] env[65121]: ERROR oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271e88b-434f-414f-2a98-2cff23d7bfe9/disk-0.vmdk due to incomplete transfer. [ 1031.659476] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bd4fdcfd-a5a0-4105-93d2-c1e2535af2fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.663487] env[65121]: WARNING nova.network.neutron [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] 5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd already exists in list: networks containing: ['5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd']. ignoring it [ 1031.673125] env[65121]: DEBUG oslo_vmware.rw_handles [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5271e88b-434f-414f-2a98-2cff23d7bfe9/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1031.673402] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Uploaded image 3cc49a60-ee19-44c9-8abb-844dc06b8184 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1031.676037] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1031.676370] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bf7c9bb4-d129-4b82-a052-994f6ac478c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.683708] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1031.683708] env[65121]: value = "task-5107050" [ 1031.683708] env[65121]: _type = "Task" [ 1031.683708] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.693278] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107050, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.706785] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.707247] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.724072] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107049, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093449} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.724346] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.725204] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f553e8c7-3900-40c6-966d-b4e744b47b5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.749099] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.753803] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41bf4324-9fbc-4d48-b038-1536c1dde14a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.777583] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1031.777583] env[65121]: value = "task-5107051" [ 1031.777583] env[65121]: _type = "Task" [ 1031.777583] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.786848] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107051, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.792822] env[65121]: WARNING neutronclient.v2_0.client [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.793578] env[65121]: WARNING openstack [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.793800] env[65121]: WARNING openstack [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.813379] env[65121]: WARNING neutronclient.v2_0.client [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.814113] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.814686] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.825030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af861e4e-1293-47f8-9487-d5a3ea94844d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.127s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.830771] env[65121]: DEBUG nova.network.neutron [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1031.839436] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.839709] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Processing image 91345b02-0e33-4bb1-bc26-4933119dd392 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.839933] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392/91345b02-0e33-4bb1-bc26-4933119dd392.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.840104] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392/91345b02-0e33-4bb1-bc26-4933119dd392.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.840291] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.840559] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0998b965-6d17-4b55-8dfb-ccb02755899b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.853651] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.853914] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.856802] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11a8a506-591f-4dec-a981-5ce42c70ea9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.864084] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1031.864084] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52559843-8cef-4765-dfe7-561bc965c6e1" [ 1031.864084] env[65121]: _type = "Task" [ 1031.864084] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.874017] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52559843-8cef-4765-dfe7-561bc965c6e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.886141] env[65121]: INFO nova.compute.manager [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Took 50.07 seconds to build instance. [ 1031.898787] env[65121]: DEBUG nova.network.neutron [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.950919] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.951386] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.020268] env[65121]: WARNING neutronclient.v2_0.client [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.021019] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.021535] env[65121]: WARNING openstack [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.085807] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.086444] env[65121]: DEBUG nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1032.089533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.342s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.089773] env[65121]: DEBUG nova.objects.instance [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lazy-loading 'resources' on Instance uuid 17f34d32-e690-40d8-99a4-acf53fa859c6 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.108046] env[65121]: DEBUG nova.network.neutron [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15f422aa-08f5-4473-b1ac-92eb03870e58", "address": "fa:16:3e:3d:ba:33", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15f422aa-08", "ovs_interfaceid": "15f422aa-08f5-4473-b1ac-92eb03870e58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1032.195016] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107050, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.290338] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107051, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.354527] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "fc3109cb-72bb-4695-82c3-a323e619919d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.354817] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "fc3109cb-72bb-4695-82c3-a323e619919d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.376786] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Preparing fetch location {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1032.377051] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Fetch image to [datastore1] OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43/OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43.vmdk {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1032.377237] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Downloading stream optimized image 91345b02-0e33-4bb1-bc26-4933119dd392 to [datastore1] OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43/OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43.vmdk on the data store datastore1 as vApp {{(pid=65121) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1032.377434] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Downloading image file data 91345b02-0e33-4bb1-bc26-4933119dd392 to the ESX as VM named 'OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43' {{(pid=65121) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1032.388758] env[65121]: DEBUG oslo_concurrency.lockutils [None req-cc3c9c20-7189-486c-b57c-3bbe7826cacf tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.578s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.420161] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Releasing lock "refresh_cache-bad70910-58a1-4e35-8d0c-3bd7c3a30abf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.420581] env[65121]: DEBUG nova.compute.manager [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1032.420835] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.422647] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34abb920-96b1-4f41-9311-769632ad5a0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.427791] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "e756fdb2-f901-4e23-92e7-f2753fd5f728" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.428037] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.428237] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "e756fdb2-f901-4e23-92e7-f2753fd5f728-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.428412] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.428599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.430746] env[65121]: INFO nova.compute.manager [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Terminating instance [ 1032.435456] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.437917] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7ded76d-2368-4866-aebe-50ec84a5b856 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.444984] env[65121]: DEBUG oslo_vmware.api [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1032.444984] env[65121]: value = "task-5107052" [ 1032.444984] env[65121]: _type = "Task" [ 1032.444984] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.455384] env[65121]: DEBUG oslo_vmware.api [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.465026] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1032.465026] env[65121]: value = "resgroup-9" [ 1032.465026] env[65121]: _type = "ResourcePool" [ 1032.465026] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1032.465268] env[65121]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-bdb50b00-bd4d-4c67-be18-55c1c416d788 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.488972] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease: (returnval){ [ 1032.488972] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52804cd7-a9ac-175e-7540-e11720ccce99" [ 1032.488972] env[65121]: _type = "HttpNfcLease" [ 1032.488972] env[65121]: } obtained for vApp import into resource pool (val){ [ 1032.488972] env[65121]: value = "resgroup-9" [ 1032.488972] env[65121]: _type = "ResourcePool" [ 1032.488972] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1032.489422] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the lease: (returnval){ [ 1032.489422] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52804cd7-a9ac-175e-7540-e11720ccce99" [ 1032.489422] env[65121]: _type = "HttpNfcLease" [ 1032.489422] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1032.497219] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1032.497219] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52804cd7-a9ac-175e-7540-e11720ccce99" [ 1032.497219] env[65121]: _type = "HttpNfcLease" [ 1032.497219] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1032.593849] env[65121]: DEBUG nova.compute.utils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1032.595883] env[65121]: DEBUG nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1032.612820] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.613480] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.613634] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.614801] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752c75ec-e7eb-43ed-b067-f42a3e6fd72e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.637023] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1032.637023] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.637023] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1032.637023] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.637342] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1032.637342] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1032.637479] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1032.637575] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1032.637729] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1032.638313] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1032.638313] env[65121]: DEBUG nova.virt.hardware [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1032.645035] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfiguring VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1032.647789] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0138c016-089e-4be5-9e1c-6934b1fab2e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.666132] env[65121]: DEBUG oslo_vmware.api [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1032.666132] env[65121]: value = "task-5107054" [ 1032.666132] env[65121]: _type = "Task" [ 1032.666132] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.680126] env[65121]: DEBUG oslo_vmware.api [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107054, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.697285] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107050, 'name': Destroy_Task, 'duration_secs': 0.908743} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.697285] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Destroyed the VM [ 1032.697285] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1032.697533] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fdd18e71-68d6-4741-8c93-135b1edf7c57 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.708760] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1032.708760] env[65121]: value = "task-5107055" [ 1032.708760] env[65121]: _type = "Task" [ 1032.708760] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.718994] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.789691] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107051, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.857968] env[65121]: DEBUG nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1032.938195] env[65121]: DEBUG nova.compute.manager [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1032.938195] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.938653] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35debc85-30cb-4558-87ee-1e68c48a903a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.954105] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.954785] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06f61797-76b3-4bc7-bd63-b7b775ddc871 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.963053] env[65121]: DEBUG oslo_vmware.api [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107052, 'name': PowerOffVM_Task, 'duration_secs': 0.155988} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.968538] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.968973] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1032.969675] env[65121]: DEBUG oslo_vmware.api [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1032.969675] env[65121]: value = "task-5107056" [ 1032.969675] env[65121]: _type = "Task" [ 1032.969675] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.970957] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80b7b1a2-0fe2-40ba-8a45-4bf3739ac4fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.993591] env[65121]: DEBUG oslo_vmware.api [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107056, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.009877] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1033.009877] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52804cd7-a9ac-175e-7540-e11720ccce99" [ 1033.009877] env[65121]: _type = "HttpNfcLease" [ 1033.009877] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1033.012121] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1033.012521] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1033.012853] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Deleting the datastore file [datastore1] bad70910-58a1-4e35-8d0c-3bd7c3a30abf {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.013249] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74b1fa2a-60a6-4ee0-b40f-e273b287cb71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.025822] env[65121]: DEBUG oslo_vmware.api [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for the task: (returnval){ [ 1033.025822] env[65121]: value = "task-5107058" [ 1033.025822] env[65121]: _type = "Task" [ 1033.025822] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.043395] env[65121]: DEBUG oslo_vmware.api [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107058, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.070666] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4540936b-950a-4da4-b257-e713f98ec80f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.079748] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7142596e-9b94-48df-abc0-127899d8576b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.115021] env[65121]: DEBUG nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1033.118698] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5942e711-d2f3-437e-a1ec-19b1fc29d24a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.129601] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14edf55-e0b8-4774-95f0-c407dfe57b1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.148702] env[65121]: DEBUG nova.compute.provider_tree [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.178337] env[65121]: DEBUG oslo_vmware.api [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107054, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.221141] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107055, 'name': RemoveSnapshot_Task, 'duration_secs': 0.476381} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.221440] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1033.221831] env[65121]: DEBUG nova.compute.manager [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1033.222740] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce75397-54bc-4ba2-90d7-01112122af81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.290125] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107051, 'name': ReconfigVM_Task, 'duration_secs': 1.252596} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.290492] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.292039] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e877449-52dc-4ef0-94ea-db9eb2f64bad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.300112] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1033.300112] env[65121]: value = "task-5107059" [ 1033.300112] env[65121]: _type = "Task" [ 1033.300112] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.311336] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107059, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.385448] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1033.486424] env[65121]: DEBUG oslo_vmware.api [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107056, 'name': PowerOffVM_Task, 'duration_secs': 0.281158} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.486770] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1033.487067] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1033.487463] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-551b4072-34ea-43f7-9b17-fc267dde68c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.496362] env[65121]: DEBUG nova.compute.manager [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1033.496362] env[65121]: DEBUG nova.compute.manager [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing instance network info cache due to event network-changed-71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1033.497134] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Acquiring lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.497134] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Acquired lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1033.497134] env[65121]: DEBUG nova.network.neutron [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Refreshing network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1033.515037] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1033.515037] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52804cd7-a9ac-175e-7540-e11720ccce99" [ 1033.515037] env[65121]: _type = "HttpNfcLease" [ 1033.515037] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1033.515037] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1033.515037] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52804cd7-a9ac-175e-7540-e11720ccce99" [ 1033.515037] env[65121]: _type = "HttpNfcLease" [ 1033.515037] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1033.515648] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a97fe89-3c92-48b3-b04d-4bcd11ce2028 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.525683] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52468db7-e039-0858-bf61-ae3274544b25/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1033.525960] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating HTTP connection to write to file with size = 31673344 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52468db7-e039-0858-bf61-ae3274544b25/disk-0.vmdk. {{(pid=65121) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1033.592081] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1033.592367] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1033.592577] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleting the datastore file [datastore1] e756fdb2-f901-4e23-92e7-f2753fd5f728 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.594170] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efce7a4b-1783-46e1-95f6-4f44873d9207 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.602776] env[65121]: DEBUG oslo_vmware.api [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Task: {'id': task-5107058, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126919} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.607286] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.607286] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.607286] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.607286] env[65121]: INFO nova.compute.manager [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1033.607286] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1033.607286] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f00cedf6-5815-4f32-b4fd-d14bfb0fe084 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.608405] env[65121]: DEBUG nova.compute.manager [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1033.608539] env[65121]: DEBUG nova.network.neutron [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1033.608787] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1033.609446] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1033.609714] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1033.618873] env[65121]: DEBUG oslo_vmware.api [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for the task: (returnval){ [ 1033.618873] env[65121]: value = "task-5107061" [ 1033.618873] env[65121]: _type = "Task" [ 1033.618873] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.633934] env[65121]: DEBUG oslo_vmware.api [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107061, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.636116] env[65121]: DEBUG nova.network.neutron [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1033.637027] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1033.653416] env[65121]: DEBUG nova.scheduler.client.report [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1033.678320] env[65121]: DEBUG oslo_vmware.api [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107054, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.737866] env[65121]: INFO nova.compute.manager [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Shelve offloading [ 1033.811874] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107059, 'name': Rename_Task, 'duration_secs': 0.200727} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.812205] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.812493] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2a35a00-d6a6-41a5-9ca5-647a759727f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.821349] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1033.821349] env[65121]: value = "task-5107062" [ 1033.821349] env[65121]: _type = "Task" [ 1033.821349] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.831402] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107062, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.000838] env[65121]: WARNING neutronclient.v2_0.client [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.002710] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.003259] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.129212] env[65121]: DEBUG nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1034.139022] env[65121]: DEBUG nova.network.neutron [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1034.140823] env[65121]: DEBUG oslo_vmware.api [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Task: {'id': task-5107061, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199233} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.141490] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.141714] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.142074] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.142227] env[65121]: INFO nova.compute.manager [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1034.142613] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1034.142883] env[65121]: DEBUG nova.compute.manager [-] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1034.143085] env[65121]: DEBUG nova.network.neutron [-] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1034.143407] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.144149] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.144563] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.154139] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.154621] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.163735] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.074s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.168763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.963s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1034.168763] env[65121]: DEBUG nova.objects.instance [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lazy-loading 'resources' on Instance uuid 90cb98b1-9520-4caf-b6c4-80abaa7a9221 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.177732] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1034.178034] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1034.178262] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1034.178497] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1034.178720] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1034.178893] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1034.179170] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1034.179742] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1034.179742] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1034.179742] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1034.179939] env[65121]: DEBUG nova.virt.hardware [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1034.181629] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec45f495-33f0-4897-9e8b-c1a8cb3c3a52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.191993] env[65121]: DEBUG oslo_vmware.api [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107054, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.198178] env[65121]: INFO nova.scheduler.client.report [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Deleted allocations for instance 17f34d32-e690-40d8-99a4-acf53fa859c6 [ 1034.215354] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a251289-2f45-4e9f-8d22-4ecc0e200b50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.239932] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.249825] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Creating folder: Project (c34ef03bfbb2498981bbb025b012861a). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1034.260988] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.265022] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.265404] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bf276ee6-edeb-4720-9e25-167f3a7bd086 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.269824] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73a04d8f-c270-4d76-9471-1f01f1dc99f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.273282] env[65121]: WARNING neutronclient.v2_0.client [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.274303] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.274770] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.301943] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1034.301943] env[65121]: value = "task-5107064" [ 1034.301943] env[65121]: _type = "Task" [ 1034.301943] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.306065] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Created folder: Project (c34ef03bfbb2498981bbb025b012861a) in parent group-v993268. [ 1034.306549] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Creating folder: Instances. Parent ref: group-v993528. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1034.312556] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b80b410a-82d1-4b81-96f5-3b8705ba49f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.334347] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1034.334783] env[65121]: DEBUG nova.compute.manager [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1034.345043] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc439f4-9ec2-4816-a756-c4117f50a943 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.346719] env[65121]: DEBUG nova.objects.instance [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.348929] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Created folder: Instances in parent group-v993528. [ 1034.349802] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1034.353076] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1034.353076] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-faa8f834-e15f-4651-bf39-c38b608aa642 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.389309] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107062, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.389309] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.389309] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.389309] env[65121]: DEBUG nova.network.neutron [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1034.400035] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.400035] env[65121]: value = "task-5107066" [ 1034.400035] env[65121]: _type = "Task" [ 1034.400035] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.414491] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107066, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.444807] env[65121]: DEBUG nova.network.neutron [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updated VIF entry in instance network info cache for port 71e0942f-5026-4128-ba81-16311feb9b3e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1034.445247] env[65121]: DEBUG nova.network.neutron [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [{"id": "71e0942f-5026-4128-ba81-16311feb9b3e", "address": "fa:16:3e:2b:a9:4a", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71e0942f-50", "ovs_interfaceid": "71e0942f-5026-4128-ba81-16311feb9b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1034.642452] env[65121]: INFO nova.compute.manager [-] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Took 1.03 seconds to deallocate network for instance. [ 1034.685642] env[65121]: DEBUG oslo_vmware.api [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107054, 'name': ReconfigVM_Task, 'duration_secs': 1.893193} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.688290] env[65121]: WARNING neutronclient.v2_0.client [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.688566] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.688802] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfigured VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1034.711661] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9b48a5c-6b0f-4b1c-b1b9-d0c5d2ad1c01 tempest-ImagesNegativeTestJSON-1426538924 tempest-ImagesNegativeTestJSON-1426538924-project-member] Lock "17f34d32-e690-40d8-99a4-acf53fa859c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.417s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.844367] env[65121]: DEBUG oslo_vmware.api [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107062, 'name': PowerOnVM_Task, 'duration_secs': 0.628166} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.846602] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.846844] env[65121]: INFO nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Took 9.12 seconds to spawn the instance on the hypervisor. [ 1034.847508] env[65121]: DEBUG nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1034.848310] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88f43a4-3b10-4c76-8c9e-e363ec01daf7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.862739] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.862915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.863087] env[65121]: DEBUG nova.network.neutron [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1034.863380] env[65121]: DEBUG nova.objects.instance [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'info_cache' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.897459] env[65121]: WARNING neutronclient.v2_0.client [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1034.898388] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1034.904027] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1034.941778] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107066, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.953935] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Releasing lock "refresh_cache-e0d3e3c8-6471-4345-8677-369612674769" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.953935] env[65121]: DEBUG nova.compute.manager [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-vif-plugged-15f422aa-08f5-4473-b1ac-92eb03870e58 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1034.953935] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.953935] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1034.953935] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.954381] env[65121]: DEBUG nova.compute.manager [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] No waiting events found dispatching network-vif-plugged-15f422aa-08f5-4473-b1ac-92eb03870e58 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1034.954381] env[65121]: WARNING nova.compute.manager [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received unexpected event network-vif-plugged-15f422aa-08f5-4473-b1ac-92eb03870e58 for instance with vm_state active and task_state None. [ 1034.954481] env[65121]: DEBUG nova.compute.manager [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-changed-15f422aa-08f5-4473-b1ac-92eb03870e58 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1034.954638] env[65121]: DEBUG nova.compute.manager [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing instance network info cache due to event network-changed-15f422aa-08f5-4473-b1ac-92eb03870e58. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1034.954995] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.955194] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.955355] env[65121]: DEBUG nova.network.neutron [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing network info cache for port 15f422aa-08f5-4473-b1ac-92eb03870e58 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1035.054279] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.055295] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.102016] env[65121]: DEBUG nova.network.neutron [-] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1035.121713] env[65121]: WARNING neutronclient.v2_0.client [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.122545] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.122759] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.151285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.157819] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37e6427-dffe-4d8b-abd5-edc200cff6e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.166535] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d557a038-ee9f-4fe7-9926-51b62f315b5e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.202577] env[65121]: DEBUG oslo_concurrency.lockutils [None req-99b6c27d-d011-44bc-8db8-4f5188cfee5d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.831s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.206921] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317bdf67-7c58-4fc1-a9c2-aa236ac96323 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.212378] env[65121]: DEBUG nova.network.neutron [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Updating instance_info_cache with network_info: [{"id": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "address": "fa:16:3e:a8:19:39", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bf01aba-25", "ovs_interfaceid": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1035.214785] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Completed reading data from the image iterator. {{(pid=65121) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1035.214974] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52468db7-e039-0858-bf61-ae3274544b25/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1035.217978] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65907ae9-e53d-4815-9f3b-bd6b343e918b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.224236] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc50f6d8-690d-44a5-ac03-666cbf252a4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.232011] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52468db7-e039-0858-bf61-ae3274544b25/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1035.232011] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52468db7-e039-0858-bf61-ae3274544b25/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1035.241720] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b4cb97cc-fcd6-4ffd-8704-e2a14301d422 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.244087] env[65121]: DEBUG nova.compute.provider_tree [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.371382] env[65121]: DEBUG nova.objects.base [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Object Instance<6fe86deb-0903-4769-a05c-b7d0acec0103> lazy-loaded attributes: flavor,info_cache {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1035.381303] env[65121]: INFO nova.compute.manager [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Took 37.38 seconds to build instance. [ 1035.430558] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107066, 'name': CreateVM_Task, 'duration_secs': 0.538996} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.430909] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1035.431558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.431558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1035.431807] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1035.432429] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-656aed0b-23a1-403b-9060-9cdaa3b214d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.438516] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1035.438516] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521e0907-3d3f-cf9f-9f42-9e56234aff05" [ 1035.438516] env[65121]: _type = "Task" [ 1035.438516] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.447964] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521e0907-3d3f-cf9f-9f42-9e56234aff05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.459119] env[65121]: WARNING neutronclient.v2_0.client [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.459894] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.460293] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.572594] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.576463] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.582651] env[65121]: DEBUG oslo_vmware.rw_handles [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52468db7-e039-0858-bf61-ae3274544b25/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1035.582872] env[65121]: INFO nova.virt.vmwareapi.images [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Downloaded image file data 91345b02-0e33-4bb1-bc26-4933119dd392 [ 1035.583718] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ffab1d-0981-4f27-87cc-be45d762390e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.600725] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19d3c373-f9cb-4e62-a9b7-87d4fcf145b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.607681] env[65121]: INFO nova.compute.manager [-] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Took 1.46 seconds to deallocate network for instance. [ 1035.644556] env[65121]: WARNING neutronclient.v2_0.client [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.645347] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.645690] env[65121]: WARNING openstack [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.698292] env[65121]: DEBUG nova.compute.manager [req-a62913b0-3e0e-4900-9eb5-8f79d8778df0 req-187b2d88-42c8-41c7-8677-4cdb5f7173ee service nova] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Received event network-vif-deleted-62496d8b-5753-4794-b358-9db503a52328 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1035.703572] env[65121]: INFO nova.virt.vmwareapi.images [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] The imported VM was unregistered [ 1035.706275] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Caching image {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1035.706553] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating directory with path [datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392 {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.707979] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ab3f6aa-b75d-4b6c-93b7-48f13fb791c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.722301] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1035.722745] env[65121]: WARNING neutronclient.v2_0.client [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.723431] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.723774] env[65121]: WARNING openstack [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.728825] env[65121]: WARNING neutronclient.v2_0.client [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.734568] env[65121]: DEBUG nova.network.neutron [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updated VIF entry in instance network info cache for port 15f422aa-08f5-4473-b1ac-92eb03870e58. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1035.734973] env[65121]: DEBUG nova.network.neutron [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15f422aa-08f5-4473-b1ac-92eb03870e58", "address": "fa:16:3e:3d:ba:33", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15f422aa-08", "ovs_interfaceid": "15f422aa-08f5-4473-b1ac-92eb03870e58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1035.739352] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created directory with path [datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392 {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.739530] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43/OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43.vmdk to [datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392/91345b02-0e33-4bb1-bc26-4933119dd392.vmdk. {{(pid=65121) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1035.739779] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-8cb41eba-692f-479c-abde-4059feb2804b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.747109] env[65121]: DEBUG nova.scheduler.client.report [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1035.751771] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1035.751771] env[65121]: value = "task-5107068" [ 1035.751771] env[65121]: _type = "Task" [ 1035.751771] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.760714] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.880166] env[65121]: WARNING neutronclient.v2_0.client [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.880902] env[65121]: WARNING openstack [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.881264] env[65121]: WARNING openstack [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.890922] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9a8fa49-bc53-450b-a624-bc6f36fbf1cf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.739s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.900163] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.900430] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.900686] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.900893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.901067] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.904959] env[65121]: INFO nova.compute.manager [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Terminating instance [ 1035.950422] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521e0907-3d3f-cf9f-9f42-9e56234aff05, 'name': SearchDatastore_Task, 'duration_secs': 0.029888} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.950541] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1035.950677] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.950924] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.951032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1035.951205] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.951804] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab070ce9-0508-455b-8bb8-7f5a74e912e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.967506] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.967759] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1035.968503] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-902c1950-188b-4871-ab84-dd5c980d1809 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.977736] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1035.977736] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5287af5e-b396-30a3-6d89-cf99a0ed6867" [ 1035.977736] env[65121]: _type = "Task" [ 1035.977736] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.986777] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5287af5e-b396-30a3-6d89-cf99a0ed6867, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.014160] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.015098] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77179a9-3e40-4bfe-8808-3c8de47e6da7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.023690] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.023986] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef595940-2758-4c5e-a3d6-8c693661d6db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.082211] env[65121]: WARNING openstack [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.082978] env[65121]: WARNING openstack [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.119573] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.128152] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.128549] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.128549] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleting the datastore file [datastore1] 6e969388-3238-404d-a8eb-e7b7318c4c72 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.128895] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cd856e2-06e0-4891-9b90-b9c0dc1a770d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.138465] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1036.138465] env[65121]: value = "task-5107070" [ 1036.138465] env[65121]: _type = "Task" [ 1036.138465] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.149424] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107070, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.168089] env[65121]: WARNING neutronclient.v2_0.client [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1036.168246] env[65121]: WARNING openstack [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.168687] env[65121]: WARNING openstack [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.238653] env[65121]: DEBUG oslo_concurrency.lockutils [req-f638fc5b-9ab4-4774-ac0d-e0cb3b0b6eed req-420a3f26-8f8d-443f-86d8-4aed9e080350 service nova] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.254417] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.258995] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.389s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.261903] env[65121]: INFO nova.compute.claims [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.283198] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.293080] env[65121]: INFO nova.scheduler.client.report [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Deleted allocations for instance 90cb98b1-9520-4caf-b6c4-80abaa7a9221 [ 1036.295621] env[65121]: DEBUG nova.network.neutron [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating instance_info_cache with network_info: [{"id": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "address": "fa:16:3e:0e:bf:f3", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape31cbeae-05", "ovs_interfaceid": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1036.334897] env[65121]: DEBUG nova.compute.manager [req-32167703-e3de-449a-868c-a893c201fb23 req-67cbc906-0754-42ed-9567-28ea943d81ec service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Received event network-vif-unplugged-3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1036.335519] env[65121]: DEBUG oslo_concurrency.lockutils [req-32167703-e3de-449a-868c-a893c201fb23 req-67cbc906-0754-42ed-9567-28ea943d81ec service nova] Acquiring lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.335816] env[65121]: DEBUG oslo_concurrency.lockutils [req-32167703-e3de-449a-868c-a893c201fb23 req-67cbc906-0754-42ed-9567-28ea943d81ec service nova] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.336042] env[65121]: DEBUG oslo_concurrency.lockutils [req-32167703-e3de-449a-868c-a893c201fb23 req-67cbc906-0754-42ed-9567-28ea943d81ec service nova] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.336264] env[65121]: DEBUG nova.compute.manager [req-32167703-e3de-449a-868c-a893c201fb23 req-67cbc906-0754-42ed-9567-28ea943d81ec service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] No waiting events found dispatching network-vif-unplugged-3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1036.336570] env[65121]: WARNING nova.compute.manager [req-32167703-e3de-449a-868c-a893c201fb23 req-67cbc906-0754-42ed-9567-28ea943d81ec service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Received unexpected event network-vif-unplugged-3bf01aba-25b1-44e7-a26e-8ffee646ba29 for instance with vm_state shelved and task_state shelving_offloading. [ 1036.409431] env[65121]: DEBUG nova.compute.manager [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1036.409720] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.410658] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a6278e-9133-43f4-a9c0-c19a017fe34b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.420228] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.420549] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7b81ac4-c0e8-42f1-a0dc-a1f6fc59706f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.428168] env[65121]: DEBUG oslo_vmware.api [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1036.428168] env[65121]: value = "task-5107071" [ 1036.428168] env[65121]: _type = "Task" [ 1036.428168] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.439604] env[65121]: DEBUG oslo_vmware.api [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.490618] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5287af5e-b396-30a3-6d89-cf99a0ed6867, 'name': SearchDatastore_Task, 'duration_secs': 0.025665} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.491756] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6c2d863-796d-4299-8d82-e2c07785bd40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.499891] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1036.499891] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a89cd3-5f43-116d-c354-d923a5e57f47" [ 1036.499891] env[65121]: _type = "Task" [ 1036.499891] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.511844] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a89cd3-5f43-116d-c354-d923a5e57f47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.651028] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107070, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.769075] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.800943] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.808163] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5eaa1270-90f2-4304-aed8-a6f27bb52afb tempest-ServersV294TestFqdnHostnames-1463148082 tempest-ServersV294TestFqdnHostnames-1463148082-project-member] Lock "90cb98b1-9520-4caf-b6c4-80abaa7a9221" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.293s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.940969] env[65121]: DEBUG oslo_vmware.api [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107071, 'name': PowerOffVM_Task, 'duration_secs': 0.236381} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.941295] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.941456] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.941754] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f16abc6e-ac96-4447-a11c-3d386118265a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.016689] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a89cd3-5f43-116d-c354-d923a5e57f47, 'name': SearchDatastore_Task, 'duration_secs': 0.075802} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.016901] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1037.017188] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 76e4cddb-b046-4a3e-9a91-d3ee847883cf/76e4cddb-b046-4a3e-9a91-d3ee847883cf.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1037.017506] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1440eaa-1163-4072-9e22-01d98e1f6958 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.021915] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.022156] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.022335] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleting the datastore file [datastore2] 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.022620] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d53989d9-5384-4530-970b-6315b729aebe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.026910] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1037.026910] env[65121]: value = "task-5107073" [ 1037.026910] env[65121]: _type = "Task" [ 1037.026910] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.033369] env[65121]: DEBUG oslo_vmware.api [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1037.033369] env[65121]: value = "task-5107074" [ 1037.033369] env[65121]: _type = "Task" [ 1037.033369] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.041263] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.047414] env[65121]: DEBUG oslo_vmware.api [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.151899] env[65121]: DEBUG oslo_vmware.api [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107070, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.932819} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.152084] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1037.152178] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1037.152289] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1037.178561] env[65121]: INFO nova.scheduler.client.report [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted allocations for instance 6e969388-3238-404d-a8eb-e7b7318c4c72 [ 1037.272824] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.547135] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.547135] env[65121]: DEBUG oslo_vmware.api [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.636901] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1037.637157] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1037.637583] env[65121]: DEBUG nova.objects.instance [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'flavor' on Instance uuid 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.684688] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1037.716676] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7734b78f-51e0-4c2b-8da2-87afe6fc22e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.725922] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6165e2-7583-4796-aed7-1cb20a3372af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.765507] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1303109f-81b1-43ec-bfc6-de115173da5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.774371] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.778162] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5b0719-9eec-42b1-ace2-816ab51c883b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.794627] env[65121]: DEBUG nova.compute.provider_tree [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.809064] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.809514] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12401c70-e8fe-465f-8d16-66b3f85d697f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.817828] env[65121]: DEBUG oslo_vmware.api [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1037.817828] env[65121]: value = "task-5107075" [ 1037.817828] env[65121]: _type = "Task" [ 1037.817828] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.830967] env[65121]: DEBUG oslo_vmware.api [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107075, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.047822] env[65121]: DEBUG oslo_vmware.api [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.001544} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.049658] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.049877] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.050118] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.050239] env[65121]: INFO nova.compute.manager [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1038.050531] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1038.050690] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.051030] env[65121]: DEBUG nova.compute.manager [-] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1038.051207] env[65121]: DEBUG nova.network.neutron [-] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1038.051418] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.051870] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1038.052142] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.092068] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.146437] env[65121]: WARNING neutronclient.v2_0.client [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.146437] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1038.147035] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.275640] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.298127] env[65121]: DEBUG nova.scheduler.client.report [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1038.333905] env[65121]: DEBUG oslo_vmware.api [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107075, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.543267] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.631282] env[65121]: DEBUG nova.objects.instance [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'pci_requests' on Instance uuid 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.775655] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.805976] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.806555] env[65121]: DEBUG nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1038.809765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 29.439s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.836176] env[65121]: DEBUG oslo_vmware.api [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107075, 'name': PowerOnVM_Task, 'duration_secs': 0.606677} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.837135] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.837549] env[65121]: DEBUG nova.compute.manager [None req-1707a0a2-671e-4bfe-87a5-26a6cb6b9195 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1038.839247] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0f555c-4d06-48ec-8d29-263c2996a0ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.907275] env[65121]: DEBUG nova.network.neutron [-] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1039.044717] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107073, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.137761] env[65121]: DEBUG nova.objects.base [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Object Instance<23d6f632-8918-46af-b239-08a9615dfbec> lazy-loaded attributes: flavor,pci_requests {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1039.138012] env[65121]: DEBUG nova.network.neutron [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1039.138362] env[65121]: WARNING neutronclient.v2_0.client [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.138650] env[65121]: WARNING neutronclient.v2_0.client [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.139678] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1039.139850] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1039.278896] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107068, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.264412} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.279224] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43/OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43.vmdk to [datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392/91345b02-0e33-4bb1-bc26-4933119dd392.vmdk. [ 1039.279427] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Cleaning up location [datastore1] OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1039.279578] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_90229488-1a5e-43d6-8147-69fe25851d43 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1039.279908] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a7c6dad-ba27-4fbc-b50c-95a7b49b056a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.291656] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1039.291656] env[65121]: value = "task-5107076" [ 1039.291656] env[65121]: _type = "Task" [ 1039.291656] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.302982] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.314177] env[65121]: DEBUG nova.compute.utils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1039.316260] env[65121]: DEBUG nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1039.316923] env[65121]: DEBUG nova.network.neutron [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1039.316923] env[65121]: WARNING neutronclient.v2_0.client [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.317233] env[65121]: WARNING neutronclient.v2_0.client [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1039.317895] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1039.318256] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1039.329242] env[65121]: INFO nova.compute.claims [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.360204] env[65121]: DEBUG nova.policy [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1039.409330] env[65121]: INFO nova.compute.manager [-] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Took 1.36 seconds to deallocate network for instance. [ 1039.544506] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107073, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.173693} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.544784] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 76e4cddb-b046-4a3e-9a91-d3ee847883cf/76e4cddb-b046-4a3e-9a91-d3ee847883cf.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.544999] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.545315] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2745140e-e67a-4303-8495-bc883f5dc82a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.552949] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1039.552949] env[65121]: value = "task-5107077" [ 1039.552949] env[65121]: _type = "Task" [ 1039.552949] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.561567] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.802544] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.065295} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.802943] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.803068] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392/91345b02-0e33-4bb1-bc26-4933119dd392.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1039.803424] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392/91345b02-0e33-4bb1-bc26-4933119dd392.vmdk to [datastore1] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1039.803766] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-703740ff-e6e6-4501-9949-21f4af416830 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.811237] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1039.811237] env[65121]: value = "task-5107078" [ 1039.811237] env[65121]: _type = "Task" [ 1039.811237] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.819244] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107078, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.826910] env[65121]: DEBUG nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1039.838186] env[65121]: INFO nova.compute.resource_tracker [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating resource usage from migration aeab73c9-8ca4-4e9d-82d4-d10914fed2b0 [ 1039.880630] env[65121]: DEBUG nova.policy [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f59f7f9f5074414e8e8efdc122f49fd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '59f0b5c46136457b99f0e2e993941cc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1039.918072] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.066481] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07335} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.067234] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.067568] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b7be7b-bd40-4379-809e-da4219da0ec5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.093958] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 76e4cddb-b046-4a3e-9a91-d3ee847883cf/76e4cddb-b046-4a3e-9a91-d3ee847883cf.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.099955] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86afe811-ff2a-43b2-9d7e-7b7a108f584a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.124559] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1040.124559] env[65121]: value = "task-5107079" [ 1040.124559] env[65121]: _type = "Task" [ 1040.124559] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.135373] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107079, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.271182] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94292b70-339a-4a8b-a0ac-e3b8d1a685d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.282261] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b86f9b6-5dc0-44d9-a600-767019f258b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.325377] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac3c80f-b9d9-496d-bebe-4f1fca9ed4e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.341240] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107078, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.342565] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e2b6d8-698d-42e4-a657-df5b5afaf30c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.360031] env[65121]: DEBUG nova.compute.provider_tree [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.635147] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107079, 'name': ReconfigVM_Task, 'duration_secs': 0.497435} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.635465] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 76e4cddb-b046-4a3e-9a91-d3ee847883cf/76e4cddb-b046-4a3e-9a91-d3ee847883cf.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.636265] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e675b913-8c83-4f5f-b60c-8a7f9db863be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.644255] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1040.644255] env[65121]: value = "task-5107080" [ 1040.644255] env[65121]: _type = "Task" [ 1040.644255] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.654758] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107080, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.831096] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107078, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.833555] env[65121]: DEBUG nova.network.neutron [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Successfully created port: 59929561-be86-4157-85cc-e3e04892acbc {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1040.837150] env[65121]: DEBUG nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1040.861288] env[65121]: DEBUG nova.scheduler.client.report [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1040.880150] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1040.880150] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1040.880431] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1040.880431] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1040.880509] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1040.880723] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1040.880917] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.881477] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1040.881477] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1040.881477] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1040.881601] env[65121]: DEBUG nova.virt.hardware [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1040.882513] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fad5a0-4988-4a3c-a159-7202e280b889 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.892633] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2332de5d-7d42-4df8-afcc-d6035ba0379f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.023826] env[65121]: DEBUG nova.network.neutron [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Successfully updated port: 065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1041.160357] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107080, 'name': Rename_Task, 'duration_secs': 0.19112} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.160799] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.162020] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4b63595-9ac8-4102-977b-8355b5fc37cc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.171839] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1041.171839] env[65121]: value = "task-5107081" [ 1041.171839] env[65121]: _type = "Task" [ 1041.171839] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.182349] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.232365] env[65121]: DEBUG nova.compute.manager [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Received event network-changed-3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1041.232644] env[65121]: DEBUG nova.compute.manager [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Refreshing instance network info cache due to event network-changed-3bf01aba-25b1-44e7-a26e-8ffee646ba29. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1041.232716] env[65121]: DEBUG oslo_concurrency.lockutils [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Acquiring lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.234253] env[65121]: DEBUG oslo_concurrency.lockutils [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Acquired lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1041.234253] env[65121]: DEBUG nova.network.neutron [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Refreshing network info cache for port 3bf01aba-25b1-44e7-a26e-8ffee646ba29 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1041.335034] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107078, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.370141] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.560s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.370701] env[65121]: INFO nova.compute.manager [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Migrating [ 1041.379296] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 30.018s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.527300] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.527582] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1041.527853] env[65121]: DEBUG nova.network.neutron [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1041.683053] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107081, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.735793] env[65121]: WARNING neutronclient.v2_0.client [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.736688] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.737098] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.833270] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107078, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.895740] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.895949] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1041.896109] env[65121]: DEBUG nova.network.neutron [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1042.032981] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.033594] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.185042] env[65121]: DEBUG oslo_vmware.api [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107081, 'name': PowerOnVM_Task, 'duration_secs': 0.715001} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.185594] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.185716] env[65121]: INFO nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Took 8.06 seconds to spawn the instance on the hypervisor. [ 1042.185871] env[65121]: DEBUG nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1042.186851] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948e754b-9603-4c6b-bc7f-729f027a102c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.333813] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107078, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.398573] env[65121]: WARNING neutronclient.v2_0.client [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.399445] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.399867] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.408505] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Applying migration context for instance 564159fe-6ecd-4276-8d65-cbe25859493a as it has an incoming, in-progress migration aeab73c9-8ca4-4e9d-82d4-d10914fed2b0. Migration status is pre-migrating {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1042.410383] env[65121]: INFO nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating resource usage from migration aeab73c9-8ca4-4e9d-82d4-d10914fed2b0 [ 1042.435573] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance c4096314-270a-4270-9e1d-5ace8ddbd286 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.435774] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 27940143-16b5-4263-b23c-354ed8ea8866 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.435905] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f84d7aa1-a90b-4c95-9673-fb1a2af7187c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.436072] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4a72fedd-b114-468e-8f34-0caec6ce73fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.436221] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 274d0ccd-c707-4a68-b280-16de2bc74d73 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.436337] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.436497] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 6fe86deb-0903-4769-a05c-b7d0acec0103 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.436647] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 59ceb0fd-1fb5-4c90-963d-fe76b9740d29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.436762] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 23d6f632-8918-46af-b239-08a9615dfbec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.436871] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance e0d3e3c8-6471-4345-8677-369612674769 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.436978] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance efb98d37-4162-4249-9f85-008d4537db87 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.437152] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance bad70910-58a1-4e35-8d0c-3bd7c3a30abf is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.437233] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance e756fdb2-f901-4e23-92e7-f2753fd5f728 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.437348] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.437483] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance ec463cad-4c80-4636-bc7a-9ec298a07d96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.438182] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance abd87ca7-5d95-4b22-8d2a-1f37abd18f86 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.438420] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.438560] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 76e4cddb-b046-4a3e-9a91-d3ee847883cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.438686] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 11caa171-0314-4e7d-8502-80932b469da8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.651086] env[65121]: WARNING nova.network.neutron [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] 5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd already exists in list: networks containing: ['5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd']. ignoring it [ 1042.651086] env[65121]: WARNING nova.network.neutron [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] 5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd already exists in list: networks containing: ['5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd']. ignoring it [ 1042.706130] env[65121]: INFO nova.compute.manager [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Took 39.40 seconds to build instance. [ 1042.800341] env[65121]: DEBUG nova.network.neutron [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Successfully updated port: 59929561-be86-4157-85cc-e3e04892acbc {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1042.834877] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107078, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.938146} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.835216] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/91345b02-0e33-4bb1-bc26-4933119dd392/91345b02-0e33-4bb1-bc26-4933119dd392.vmdk to [datastore1] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1042.836037] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6494a369-985e-4460-aa62-a9eb349a620f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.871887] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.872327] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53f20b57-a127-47b5-9a91-5a828c9f95d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.897205] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1042.897205] env[65121]: value = "task-5107082" [ 1042.897205] env[65121]: _type = "Task" [ 1042.897205] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.909706] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.941908] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 84ae6061-a962-4a74-8661-9718cc4c5346 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1042.942165] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Migration aeab73c9-8ca4-4e9d-82d4-d10914fed2b0 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1042.942399] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 564159fe-6ecd-4276-8d65-cbe25859493a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1042.967936] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.968570] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.031328] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.031706] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.069219] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.070099] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.134195] env[65121]: WARNING neutronclient.v2_0.client [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.136037] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.136656] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.195295] env[65121]: WARNING neutronclient.v2_0.client [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.195968] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.196333] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.211912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f0242b2-3530-4d9e-8767-58262f7a8430 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.277s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1043.296220] env[65121]: WARNING neutronclient.v2_0.client [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.297137] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.297546] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.306824] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "refresh_cache-11caa171-0314-4e7d-8502-80932b469da8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.307064] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquired lock "refresh_cache-11caa171-0314-4e7d-8502-80932b469da8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.307210] env[65121]: DEBUG nova.network.neutron [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1043.411515] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107082, 'name': ReconfigVM_Task, 'duration_secs': 0.316427} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.411916] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfigured VM instance instance-0000003a to attach disk [datastore1] e0d3e3c8-6471-4345-8677-369612674769/e0d3e3c8-6471-4345-8677-369612674769.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.413487] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'boot_index': 0, 'encrypted': False, 'encryption_format': None, 'encryption_options': None, 'device_type': 'disk', 'disk_bus': None, 'encryption_secret_uuid': None, 'guest_format': None, 'size': 0, 'image_id': '3ed5d656-a616-4609-8bb7-b02beb3ac3df'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'attachment_id': 'bd51ab0a-aead-4779-9adc-0db1b6b8572c', 'device_type': None, 'delete_on_termination': False, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993500', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'name': 'volume-02de1e93-4004-47b3-a1b4-d2690ef0da94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'e0d3e3c8-6471-4345-8677-369612674769', 'attached_at': '', 'detached_at': '', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'serial': '02de1e93-4004-47b3-a1b4-d2690ef0da94'}, 'mount_device': '/dev/sdb', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65121) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1043.413735] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1043.414041] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993500', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'name': 'volume-02de1e93-4004-47b3-a1b4-d2690ef0da94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'e0d3e3c8-6471-4345-8677-369612674769', 'attached_at': '', 'detached_at': '', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'serial': '02de1e93-4004-47b3-a1b4-d2690ef0da94'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1043.415910] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2261659-d9f0-45a5-a923-fa5594c53d2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.433633] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af0c7bc-7788-4046-923f-77909bc74139 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.453293] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 3d6fc9df-6872-4929-8813-bec1db1ebb16 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1043.463314] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] volume-02de1e93-4004-47b3-a1b4-d2690ef0da94/volume-02de1e93-4004-47b3-a1b4-d2690ef0da94.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.468434] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-006f137e-891a-45b1-a38c-ccbeb944820a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.489722] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1043.489722] env[65121]: value = "task-5107083" [ 1043.489722] env[65121]: _type = "Task" [ 1043.489722] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.491876] env[65121]: DEBUG nova.network.neutron [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Updated VIF entry in instance network info cache for port 3bf01aba-25b1-44e7-a26e-8ffee646ba29. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1043.492306] env[65121]: DEBUG nova.network.neutron [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Updating instance_info_cache with network_info: [{"id": "3bf01aba-25b1-44e7-a26e-8ffee646ba29", "address": "fa:16:3e:a8:19:39", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": null, "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap3bf01aba-25", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.504150] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107083, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.506327] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.506776] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.587730] env[65121]: DEBUG nova.network.neutron [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance_info_cache with network_info: [{"id": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "address": "fa:16:3e:d9:2e:bc", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb0cba48-a3", "ovs_interfaceid": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.662415] env[65121]: WARNING neutronclient.v2_0.client [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.663152] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.663514] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.782849] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.783287] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.819760] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.820168] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.856285] env[65121]: DEBUG nova.network.neutron [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1043.862128] env[65121]: WARNING neutronclient.v2_0.client [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.862760] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.863103] env[65121]: WARNING openstack [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.879533] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.880044] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.943664] env[65121]: WARNING neutronclient.v2_0.client [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.944328] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.944735] env[65121]: WARNING openstack [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.964959] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance fc3109cb-72bb-4695-82c3-a323e619919d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1043.965266] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1043.965426] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3648MB phys_disk=100GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '15', 'num_vm_active': '12', 'num_task_None': '11', 'num_os_type_None': '15', 'num_proj_72237a0d762645588c41231b0a34a796': '1', 'io_workload': '3', 'num_proj_89110618b6ef4ce68b2ca53e7e168139': '2', 'num_proj_00383aa3355e438cb703c2b86c7917f2': '2', 'num_vm_shelved_offloaded': '1', 'num_task_spawning': '3', 'num_proj_ad1aea30d62c45e193c5a54c429ce7ba': '1', 'num_proj_48809b7bf36c4315b9ac315d6b9ca0c8': '1', 'num_proj_1ce7f6698e214d73ae43427601058af8': '1', 'num_proj_f056059180af48bba4587006efb0b211': '1', 'num_proj_35312a302644426f98f127e89a067e75': '1', 'num_task_resize_prep': '1', 'num_proj_e491b88dcaad42dcb345b3217bc6f941': '1', 'num_proj_4feb04f47a97427c8a2bc97883332e1d': '2', 'num_vm_building': '2', 'num_proj_c34ef03bfbb2498981bbb025b012861a': '1', 'num_proj_59f0b5c46136457b99f0e2e993941cc0': '1'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1043.979304] env[65121]: DEBUG nova.network.neutron [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15f422aa-08f5-4473-b1ac-92eb03870e58", "address": "fa:16:3e:3d:ba:33", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15f422aa-08", "ovs_interfaceid": "15f422aa-08f5-4473-b1ac-92eb03870e58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "065401c7-8dd5-4be7-a65a-d1c11a0d2ba6", "address": "fa:16:3e:45:52:fa", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap065401c7-8d", "ovs_interfaceid": "065401c7-8dd5-4be7-a65a-d1c11a0d2ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1043.994726] env[65121]: DEBUG oslo_concurrency.lockutils [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Releasing lock "refresh_cache-6e969388-3238-404d-a8eb-e7b7318c4c72" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.995331] env[65121]: DEBUG nova.compute.manager [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Received event network-changed-53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1043.995331] env[65121]: DEBUG nova.compute.manager [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Refreshing instance network info cache due to event network-changed-53616727-ab6a-4f06-a597-0fc068a18955. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1043.995468] env[65121]: DEBUG oslo_concurrency.lockutils [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Acquiring lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.995520] env[65121]: DEBUG oslo_concurrency.lockutils [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Acquired lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1043.995646] env[65121]: DEBUG nova.network.neutron [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Refreshing network info cache for port 53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1044.014212] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107083, 'name': ReconfigVM_Task, 'duration_secs': 0.345428} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.014567] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfigured VM instance instance-0000003a to attach disk [datastore1] volume-02de1e93-4004-47b3-a1b4-d2690ef0da94/volume-02de1e93-4004-47b3-a1b4-d2690ef0da94.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1044.027667] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0522ddc0-9d55-47f5-983e-0d70db44b6af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.047220] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1044.047220] env[65121]: value = "task-5107084" [ 1044.047220] env[65121]: _type = "Task" [ 1044.047220] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.057440] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107084, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.091893] env[65121]: DEBUG nova.network.neutron [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Updating instance_info_cache with network_info: [{"id": "59929561-be86-4157-85cc-e3e04892acbc", "address": "fa:16:3e:ab:4c:3b", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59929561-be", "ovs_interfaceid": "59929561-be86-4157-85cc-e3e04892acbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1044.093474] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.353346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "6e969388-3238-404d-a8eb-e7b7318c4c72" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.382305] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c95f723-7e25-40c8-9105-73e156d203dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.390995] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a84c3a-5cbb-4276-ae8f-649fe2b7de83 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.423752] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b0cea9-dfbb-438c-bb78-4f4c9bab6043 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.432521] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d5b340-395f-4855-b53c-9fddbbc3f3ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.447143] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1044.481928] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.482724] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.482873] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1044.483742] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b402bb-4c23-43b3-8d05-1baf22802fe5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.503501] env[65121]: WARNING neutronclient.v2_0.client [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1044.504179] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.504513] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.513274] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1044.513592] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1044.513812] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1044.514068] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1044.514254] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1044.514471] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1044.514747] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.515063] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1044.515222] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1044.515446] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1044.515780] env[65121]: DEBUG nova.virt.hardware [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1044.523032] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfiguring VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1044.523633] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f09a828-2ffa-4917-91ab-7b9243cb1d60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.542454] env[65121]: DEBUG oslo_vmware.api [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1044.542454] env[65121]: value = "task-5107085" [ 1044.542454] env[65121]: _type = "Task" [ 1044.542454] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.552191] env[65121]: DEBUG oslo_vmware.api [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107085, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.560761] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107084, 'name': ReconfigVM_Task, 'duration_secs': 0.163719} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.563693] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993500', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'name': 'volume-02de1e93-4004-47b3-a1b4-d2690ef0da94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'e0d3e3c8-6471-4345-8677-369612674769', 'attached_at': '', 'detached_at': '', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'serial': '02de1e93-4004-47b3-a1b4-d2690ef0da94'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1044.564832] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e165a00-8fce-49e2-90d1-ed3adde9f9fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.571033] env[65121]: DEBUG nova.compute.manager [req-4de0fd0c-323e-4bff-b2ed-692b04fe054f req-35567056-48e0-4ed4-a6cc-b42c647e4bc6 service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Received event network-vif-plugged-59929561-be86-4157-85cc-e3e04892acbc {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1044.571033] env[65121]: DEBUG oslo_concurrency.lockutils [req-4de0fd0c-323e-4bff-b2ed-692b04fe054f req-35567056-48e0-4ed4-a6cc-b42c647e4bc6 service nova] Acquiring lock "11caa171-0314-4e7d-8502-80932b469da8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.571274] env[65121]: DEBUG oslo_concurrency.lockutils [req-4de0fd0c-323e-4bff-b2ed-692b04fe054f req-35567056-48e0-4ed4-a6cc-b42c647e4bc6 service nova] Lock "11caa171-0314-4e7d-8502-80932b469da8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.571442] env[65121]: DEBUG oslo_concurrency.lockutils [req-4de0fd0c-323e-4bff-b2ed-692b04fe054f req-35567056-48e0-4ed4-a6cc-b42c647e4bc6 service nova] Lock "11caa171-0314-4e7d-8502-80932b469da8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.571442] env[65121]: DEBUG nova.compute.manager [req-4de0fd0c-323e-4bff-b2ed-692b04fe054f req-35567056-48e0-4ed4-a6cc-b42c647e4bc6 service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] No waiting events found dispatching network-vif-plugged-59929561-be86-4157-85cc-e3e04892acbc {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1044.572095] env[65121]: WARNING nova.compute.manager [req-4de0fd0c-323e-4bff-b2ed-692b04fe054f req-35567056-48e0-4ed4-a6cc-b42c647e4bc6 service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Received unexpected event network-vif-plugged-59929561-be86-4157-85cc-e3e04892acbc for instance with vm_state building and task_state spawning. [ 1044.573555] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1044.573555] env[65121]: value = "task-5107086" [ 1044.573555] env[65121]: _type = "Task" [ 1044.573555] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.583700] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107086, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.595910] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Releasing lock "refresh_cache-11caa171-0314-4e7d-8502-80932b469da8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1044.596539] env[65121]: DEBUG nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Instance network_info: |[{"id": "59929561-be86-4157-85cc-e3e04892acbc", "address": "fa:16:3e:ab:4c:3b", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59929561-be", "ovs_interfaceid": "59929561-be86-4157-85cc-e3e04892acbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1044.596783] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:4c:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac2c9d07-ed01-47a9-88f1-562992bc1076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59929561-be86-4157-85cc-e3e04892acbc', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.604511] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Creating folder: Project (59f0b5c46136457b99f0e2e993941cc0). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.613776] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb8c04a1-5b52-49cc-9328-740bf5630453 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.632431] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Created folder: Project (59f0b5c46136457b99f0e2e993941cc0) in parent group-v993268. [ 1044.632431] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Creating folder: Instances. Parent ref: group-v993531. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1044.632431] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62be8f7c-7104-439f-aaab-6be96fb8d968 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.643308] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Created folder: Instances in parent group-v993531. [ 1044.643604] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1044.643822] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.644055] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84f0eefd-257c-4428-9e3a-66cf4169a4cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.668052] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1044.668052] env[65121]: value = "task-5107089" [ 1044.668052] env[65121]: _type = "Task" [ 1044.668052] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.673822] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.674223] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.683113] env[65121]: DEBUG nova.compute.manager [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Received event network-vif-deleted-acee4131-75b4-4eef-9802-d42c3806fa26 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1044.683398] env[65121]: DEBUG nova.compute.manager [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-vif-plugged-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1044.683654] env[65121]: DEBUG oslo_concurrency.lockutils [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.684778] env[65121]: DEBUG oslo_concurrency.lockutils [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.684991] env[65121]: DEBUG oslo_concurrency.lockutils [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.685201] env[65121]: DEBUG nova.compute.manager [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] No waiting events found dispatching network-vif-plugged-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1044.685519] env[65121]: WARNING nova.compute.manager [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received unexpected event network-vif-plugged-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 for instance with vm_state active and task_state None. [ 1044.686908] env[65121]: DEBUG nova.compute.manager [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-changed-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1044.687493] env[65121]: DEBUG nova.compute.manager [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing instance network info cache due to event network-changed-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1044.687493] env[65121]: DEBUG oslo_concurrency.lockutils [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.687493] env[65121]: DEBUG oslo_concurrency.lockutils [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1044.687649] env[65121]: DEBUG nova.network.neutron [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Refreshing network info cache for port 065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1044.698070] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107089, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.758990] env[65121]: WARNING neutronclient.v2_0.client [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1044.759701] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1044.760159] env[65121]: WARNING openstack [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.851281] env[65121]: DEBUG nova.network.neutron [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updated VIF entry in instance network info cache for port 53616727-ab6a-4f06-a597-0fc068a18955. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1044.851641] env[65121]: DEBUG nova.network.neutron [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updating instance_info_cache with network_info: [{"id": "53616727-ab6a-4f06-a597-0fc068a18955", "address": "fa:16:3e:90:9a:47", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap53616727-ab", "ovs_interfaceid": "53616727-ab6a-4f06-a597-0fc068a18955", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1044.974957] env[65121]: ERROR nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [req-583c0fcf-b88f-4ed0-bd40-95d3c82af7d9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-583c0fcf-b88f-4ed0-bd40-95d3c82af7d9"}]} [ 1044.992105] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1045.008261] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1045.008261] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.024310] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1045.048405] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1045.057098] env[65121]: DEBUG oslo_vmware.api [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.083924] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107086, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.185358] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107089, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.193065] env[65121]: WARNING neutronclient.v2_0.client [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.193804] env[65121]: WARNING openstack [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1045.194243] env[65121]: WARNING openstack [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1045.355453] env[65121]: WARNING openstack [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1045.356083] env[65121]: WARNING openstack [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1045.363811] env[65121]: DEBUG oslo_concurrency.lockutils [req-d2e18ce7-64be-4aa4-9442-35d33b418922 req-78ce0e83-5cf7-4587-b624-84ac6106277f service nova] Releasing lock "refresh_cache-1fbd3a12-82d3-458d-a590-0f5c0c70ae55" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1045.449532] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2161e106-21eb-4d8a-b0e6-d4c25540252c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.455181] env[65121]: WARNING neutronclient.v2_0.client [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.455956] env[65121]: WARNING openstack [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1045.456489] env[65121]: WARNING openstack [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1045.469356] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263bbe1f-6570-4048-aa92-ad43ee9511ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.506815] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db545cc-0235-4512-bd0c-75fba4a844ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.515989] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c47d040-c423-4a6c-afcc-b640a1cb9976 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.536746] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.554553] env[65121]: DEBUG oslo_vmware.api [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107085, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.564896] env[65121]: DEBUG nova.network.neutron [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updated VIF entry in instance network info cache for port 065401c7-8dd5-4be7-a65a-d1c11a0d2ba6. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1045.565464] env[65121]: DEBUG nova.network.neutron [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15f422aa-08f5-4473-b1ac-92eb03870e58", "address": "fa:16:3e:3d:ba:33", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15f422aa-08", "ovs_interfaceid": "15f422aa-08f5-4473-b1ac-92eb03870e58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "065401c7-8dd5-4be7-a65a-d1c11a0d2ba6", "address": "fa:16:3e:45:52:fa", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap065401c7-8d", "ovs_interfaceid": "065401c7-8dd5-4be7-a65a-d1c11a0d2ba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1045.585278] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107086, 'name': Rename_Task, 'duration_secs': 0.895407} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.586176] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1045.586432] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00db0195-f719-4e86-9e59-9edb6e9cd85b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.594252] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1045.594252] env[65121]: value = "task-5107090" [ 1045.594252] env[65121]: _type = "Task" [ 1045.594252] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.606022] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.629813] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e96495b-37ba-44c4-8d41-21269062d773 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.650611] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance '564159fe-6ecd-4276-8d65-cbe25859493a' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1045.686685] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107089, 'name': CreateVM_Task, 'duration_secs': 0.762662} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.686832] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.687777] env[65121]: WARNING neutronclient.v2_0.client [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.687909] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.688110] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.688435] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1045.688711] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f34b9ca-a98e-4770-b24a-0a277200a1d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.697686] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1045.697686] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab8664-a674-05e8-d9f1-f3ef8321b36a" [ 1045.697686] env[65121]: _type = "Task" [ 1045.697686] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.707750] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ab8664-a674-05e8-d9f1-f3ef8321b36a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.839488] env[65121]: DEBUG nova.compute.manager [None req-183e0c19-b0c2-4743-9aba-93e2395ce095 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1045.840683] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c420d8ce-5f1d-4847-acab-9b4bb305b72c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.054747] env[65121]: DEBUG oslo_vmware.api [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107085, 'name': ReconfigVM_Task, 'duration_secs': 1.434272} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.055265] env[65121]: WARNING neutronclient.v2_0.client [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.055563] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1046.055778] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfigured VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1046.068673] env[65121]: DEBUG oslo_concurrency.lockutils [req-845be60f-fcfb-46b8-bc20-33efbe890a93 req-33e95adf-0919-4fdd-bced-4152ee50db0d service nova] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1046.073866] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 125 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1046.074139] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 125 to 126 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1046.074298] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1046.106473] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107090, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.157782] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.158122] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34ac6e44-0403-454c-bf34-df6ea6a11726 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.167627] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1046.167627] env[65121]: value = "task-5107091" [ 1046.167627] env[65121]: _type = "Task" [ 1046.167627] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.178085] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107091, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.209607] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ab8664-a674-05e8-d9f1-f3ef8321b36a, 'name': SearchDatastore_Task, 'duration_secs': 0.019161} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.210331] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1046.210536] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.210766] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.210904] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1046.211285] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.211370] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d06e7c4-2f41-483c-9ab8-9bb7f717bd30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.222150] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.222464] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1046.223512] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ed66ab7-bce0-496d-a7cc-ec96e2d8ae2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.232313] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1046.232313] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52829b71-c358-cd3f-b847-3025139e44b0" [ 1046.232313] env[65121]: _type = "Task" [ 1046.232313] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.243740] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52829b71-c358-cd3f-b847-3025139e44b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.355482] env[65121]: INFO nova.compute.manager [None req-183e0c19-b0c2-4743-9aba-93e2395ce095 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] instance snapshotting [ 1046.356144] env[65121]: DEBUG nova.objects.instance [None req-183e0c19-b0c2-4743-9aba-93e2395ce095 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lazy-loading 'flavor' on Instance uuid 76e4cddb-b046-4a3e-9a91-d3ee847883cf {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.560887] env[65121]: DEBUG oslo_concurrency.lockutils [None req-00c8a74d-5de6-4d30-96d6-d154c69f5dbb tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.924s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.579444] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1046.579683] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.201s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.579933] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.585s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.581491] env[65121]: INFO nova.compute.claims [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.605731] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107090, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.677637] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107091, 'name': PowerOffVM_Task, 'duration_secs': 0.345912} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.677942] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1046.678135] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance '564159fe-6ecd-4276-8d65-cbe25859493a' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1046.743290] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52829b71-c358-cd3f-b847-3025139e44b0, 'name': SearchDatastore_Task, 'duration_secs': 0.013318} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.744125] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5021dc17-3a3b-403d-b3d9-932abea8b687 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.751193] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1046.751193] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dec065-ddef-f588-4186-d27e6475e181" [ 1046.751193] env[65121]: _type = "Task" [ 1046.751193] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.759713] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dec065-ddef-f588-4186-d27e6475e181, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.808345] env[65121]: DEBUG nova.compute.manager [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Received event network-changed-59929561-be86-4157-85cc-e3e04892acbc {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1046.808690] env[65121]: DEBUG nova.compute.manager [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Refreshing instance network info cache due to event network-changed-59929561-be86-4157-85cc-e3e04892acbc. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1046.809008] env[65121]: DEBUG oslo_concurrency.lockutils [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Acquiring lock "refresh_cache-11caa171-0314-4e7d-8502-80932b469da8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.809186] env[65121]: DEBUG oslo_concurrency.lockutils [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Acquired lock "refresh_cache-11caa171-0314-4e7d-8502-80932b469da8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1046.809401] env[65121]: DEBUG nova.network.neutron [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Refreshing network info cache for port 59929561-be86-4157-85cc-e3e04892acbc {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1046.862311] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e7d532-db33-4232-8522-476d69ddfb4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.880439] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a01d9a6-50f1-4e6d-b267-8d65c77bd95f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.934318] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.934562] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.934797] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.934992] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.935183] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.938041] env[65121]: INFO nova.compute.manager [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Terminating instance [ 1047.106715] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107090, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.185474] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1047.185474] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1047.185751] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1047.185944] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1047.186818] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1047.186818] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1047.186818] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1047.186818] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1047.187136] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1047.187136] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1047.187201] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1047.192931] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0e2565b-3d80-4e3b-a51b-6dea9e0a1439 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.209783] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1047.209783] env[65121]: value = "task-5107092" [ 1047.209783] env[65121]: _type = "Task" [ 1047.209783] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.219604] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107092, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.263557] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dec065-ddef-f588-4186-d27e6475e181, 'name': SearchDatastore_Task, 'duration_secs': 0.02243} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.263793] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.264272] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 11caa171-0314-4e7d-8502-80932b469da8/11caa171-0314-4e7d-8502-80932b469da8.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1047.264739] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f15821c7-2607-47e3-b1f3-91b51688be83 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.278409] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1047.278409] env[65121]: value = "task-5107093" [ 1047.278409] env[65121]: _type = "Task" [ 1047.278409] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.290953] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.312969] env[65121]: WARNING neutronclient.v2_0.client [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.313845] env[65121]: WARNING openstack [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.314334] env[65121]: WARNING openstack [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.391149] env[65121]: DEBUG nova.compute.manager [None req-183e0c19-b0c2-4743-9aba-93e2395ce095 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Instance disappeared during snapshot {{(pid=65121) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4652}} [ 1047.443495] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "refresh_cache-76e4cddb-b046-4a3e-9a91-d3ee847883cf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.443675] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquired lock "refresh_cache-76e4cddb-b046-4a3e-9a91-d3ee847883cf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.443888] env[65121]: DEBUG nova.network.neutron [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1047.496251] env[65121]: WARNING openstack [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.496738] env[65121]: WARNING openstack [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.539426] env[65121]: DEBUG nova.compute.manager [None req-183e0c19-b0c2-4743-9aba-93e2395ce095 tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Found 0 images (rotation: 2) {{(pid=65121) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 1047.609291] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107090, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.638169] env[65121]: WARNING neutronclient.v2_0.client [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.641023] env[65121]: WARNING openstack [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.641023] env[65121]: WARNING openstack [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.733906] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107092, 'name': ReconfigVM_Task, 'duration_secs': 0.237749} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.734273] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance '564159fe-6ecd-4276-8d65-cbe25859493a' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1047.793717] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107093, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.874884] env[65121]: DEBUG nova.network.neutron [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Updated VIF entry in instance network info cache for port 59929561-be86-4157-85cc-e3e04892acbc. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1047.876141] env[65121]: DEBUG nova.network.neutron [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Updating instance_info_cache with network_info: [{"id": "59929561-be86-4157-85cc-e3e04892acbc", "address": "fa:16:3e:ab:4c:3b", "network": {"id": "fc80040b-d829-4253-9eca-edb1e38012bf", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "009aecc07c7c45ccbe2e2a9f13e8ce6b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac2c9d07-ed01-47a9-88f1-562992bc1076", "external-id": "nsx-vlan-transportzone-968", "segmentation_id": 968, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59929561-be", "ovs_interfaceid": "59929561-be86-4157-85cc-e3e04892acbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1047.950571] env[65121]: WARNING neutronclient.v2_0.client [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.951331] env[65121]: WARNING openstack [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.951667] env[65121]: WARNING openstack [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.992795] env[65121]: DEBUG nova.network.neutron [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1048.074214] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7823c7-9c4b-431e-9842-e91d05852f97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.085391] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d438a1c6-7544-4b28-92e4-eea2d27623eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.095811] env[65121]: DEBUG nova.network.neutron [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.129966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Releasing lock "refresh_cache-76e4cddb-b046-4a3e-9a91-d3ee847883cf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.130390] env[65121]: DEBUG nova.compute.manager [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1048.130815] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.131414] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1fe8ca-7acf-43b6-91a9-bce23a17d55b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.135326] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ce2ddb-25c8-499b-b96d-e7d67368a537 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.151327] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4a9cfd-d62a-440b-9d9d-fe489c37c41b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.158894] env[65121]: DEBUG oslo_vmware.api [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107090, 'name': PowerOnVM_Task, 'duration_secs': 2.125684} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.158894] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.159406] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.161484] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5dfa37f9-579a-49a0-9917-9c3b470faa0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.175197] env[65121]: DEBUG nova.compute.provider_tree [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.178597] env[65121]: DEBUG oslo_vmware.api [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1048.178597] env[65121]: value = "task-5107094" [ 1048.178597] env[65121]: _type = "Task" [ 1048.178597] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.188271] env[65121]: DEBUG oslo_vmware.api [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107094, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.251224] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1048.251491] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1048.251720] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1048.252030] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1048.252227] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1048.252393] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1048.252625] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.252806] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1048.252997] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1048.253175] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1048.253362] env[65121]: DEBUG nova.virt.hardware [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1048.261057] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1048.261962] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4267809-da28-4bdf-8fc5-a5d9c3b1dfd0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.280568] env[65121]: DEBUG nova.compute.manager [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1048.281574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b999ac-ce8c-4fb6-809d-12b43e40239c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.287137] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1048.287137] env[65121]: value = "task-5107095" [ 1048.287137] env[65121]: _type = "Task" [ 1048.287137] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.303399] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675149} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.303956] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 11caa171-0314-4e7d-8502-80932b469da8/11caa171-0314-4e7d-8502-80932b469da8.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1048.304197] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.304466] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ffbd3365-5c2d-4f05-a4de-681a3dd9fc97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.309906] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107095, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.317728] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1048.317728] env[65121]: value = "task-5107096" [ 1048.317728] env[65121]: _type = "Task" [ 1048.317728] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.335193] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107096, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.381855] env[65121]: DEBUG oslo_concurrency.lockutils [req-5d3aa3e2-0ec3-4b22-aed9-6c899bd68c0a req-fe4759a8-83ed-47f4-ac34-301fb2108abb service nova] Releasing lock "refresh_cache-11caa171-0314-4e7d-8502-80932b469da8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.683263] env[65121]: DEBUG nova.scheduler.client.report [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1048.702886] env[65121]: DEBUG oslo_vmware.api [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107094, 'name': PowerOffVM_Task, 'duration_secs': 0.412218} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.704180] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.706981] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1048.708050] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56d52050-5495-4c5d-a09d-e5229d5fefbf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.738350] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1048.738915] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1048.739287] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Deleting the datastore file [datastore1] 76e4cddb-b046-4a3e-9a91-d3ee847883cf {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1048.739768] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c8defe1-67fa-4a3a-8f06-cbb1eda956a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.748631] env[65121]: DEBUG oslo_vmware.api [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for the task: (returnval){ [ 1048.748631] env[65121]: value = "task-5107098" [ 1048.748631] env[65121]: _type = "Task" [ 1048.748631] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.759081] env[65121]: DEBUG oslo_vmware.api [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107098, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.802776] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107095, 'name': ReconfigVM_Task, 'duration_secs': 0.392992} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.803537] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1048.808677] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3195f23-a544-49ed-a9bc-93646c4fa3a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.816917] env[65121]: DEBUG oslo_concurrency.lockutils [None req-20b53d02-eee9-4e09-b10b-6789b30cab5e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 57.828s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.839618] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 564159fe-6ecd-4276-8d65-cbe25859493a/564159fe-6ecd-4276-8d65-cbe25859493a.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.844082] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d13148c-abae-4506-936b-405c276154ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.870542] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107096, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136893} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.872833] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.876153] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1048.876153] env[65121]: value = "task-5107099" [ 1048.876153] env[65121]: _type = "Task" [ 1048.876153] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.876153] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a40244-e804-4505-ac14-b038502f3960 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.904125] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 11caa171-0314-4e7d-8502-80932b469da8/11caa171-0314-4e7d-8502-80932b469da8.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.908789] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42dc89b9-371b-47ea-9c0a-246d34752d2a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.928167] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.935886] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1048.935886] env[65121]: value = "task-5107100" [ 1048.935886] env[65121]: _type = "Task" [ 1048.935886] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.946621] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.192668] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.193116] env[65121]: DEBUG nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1049.196690] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.864s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.197052] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.201025] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.743s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.201922] env[65121]: INFO nova.compute.claims [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.248215] env[65121]: INFO nova.scheduler.client.report [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted allocations for instance 274d0ccd-c707-4a68-b280-16de2bc74d73 [ 1049.270212] env[65121]: DEBUG oslo_vmware.api [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Task: {'id': task-5107098, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209555} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.271070] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.271070] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1049.271070] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1049.272258] env[65121]: INFO nova.compute.manager [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1049.272258] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1049.272258] env[65121]: DEBUG nova.compute.manager [-] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1049.272258] env[65121]: DEBUG nova.network.neutron [-] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1049.272258] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.273184] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.273301] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.390660] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107099, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.400574] env[65121]: DEBUG nova.network.neutron [-] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1049.400843] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.448783] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107100, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.682228] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-15f422aa-08f5-4473-b1ac-92eb03870e58" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.682486] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-15f422aa-08f5-4473-b1ac-92eb03870e58" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.706820] env[65121]: DEBUG nova.compute.utils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1049.710787] env[65121]: DEBUG nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1049.711146] env[65121]: DEBUG nova.network.neutron [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1049.711477] env[65121]: WARNING neutronclient.v2_0.client [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.715190] env[65121]: WARNING neutronclient.v2_0.client [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.715190] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.715190] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.766619] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4bac191e-53a3-4339-a8c7-3cc36a03f898 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "274d0ccd-c707-4a68-b280-16de2bc74d73" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.279s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.771318] env[65121]: DEBUG nova.policy [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63bf1d576d7e44c3a8f99a75546efe52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4467da3ed41245ddbc93fc865a8b7bdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1049.894174] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107099, 'name': ReconfigVM_Task, 'duration_secs': 0.582927} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.897242] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 564159fe-6ecd-4276-8d65-cbe25859493a/564159fe-6ecd-4276-8d65-cbe25859493a.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.897399] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance '564159fe-6ecd-4276-8d65-cbe25859493a' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.903303] env[65121]: DEBUG nova.network.neutron [-] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1049.950294] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107100, 'name': ReconfigVM_Task, 'duration_secs': 0.689666} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.952506] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 11caa171-0314-4e7d-8502-80932b469da8/11caa171-0314-4e7d-8502-80932b469da8.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.952506] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fad28a6-a5ed-4a6c-a8a3-7d8d3b2af27b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.965858] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1049.965858] env[65121]: value = "task-5107101" [ 1049.965858] env[65121]: _type = "Task" [ 1049.965858] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.976279] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107101, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.140261] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adc9c40-b3f9-4f98-aa80-22979670423a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.148029] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd2f906-4314-4361-bbdd-d5d3417d42f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.183315] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94db7b4f-956d-497d-8f6b-e0c7b075fe78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.186653] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.187356] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.187760] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac79733-124f-458c-a4f1-29ff08a01d8d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.210497] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e487ea78-be3c-4afe-b05e-e18bdd8c4adb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.214717] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31f66dc-cbdf-4134-8efc-cd1254949028 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.230593] env[65121]: DEBUG nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1050.233576] env[65121]: DEBUG nova.compute.provider_tree [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1050.258140] env[65121]: WARNING neutronclient.v2_0.client [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.262845] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfiguring VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1050.267075] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b00a7047-8ca6-4b41-81f2-8d0d7801fe67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.289742] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1050.289742] env[65121]: value = "task-5107102" [ 1050.289742] env[65121]: _type = "Task" [ 1050.289742] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.301447] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.318238] env[65121]: DEBUG nova.network.neutron [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Successfully created port: 2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1050.343226] env[65121]: DEBUG nova.scheduler.client.report [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 126 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1050.343607] env[65121]: DEBUG nova.compute.provider_tree [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 126 to 127 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1050.343825] env[65121]: DEBUG nova.compute.provider_tree [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1050.405632] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cabb51-9790-4421-b6d8-a2a765030f27 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.408903] env[65121]: INFO nova.compute.manager [-] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Took 1.14 seconds to deallocate network for instance. [ 1050.432701] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6976d7-b98e-49ce-b6ee-21259b5c5432 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.455770] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance '564159fe-6ecd-4276-8d65-cbe25859493a' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1050.480279] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107101, 'name': Rename_Task, 'duration_secs': 0.175679} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.480386] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.480587] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0529f07-c5a2-4d5f-a6b1-0c851e283f69 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.489826] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1050.489826] env[65121]: value = "task-5107103" [ 1050.489826] env[65121]: _type = "Task" [ 1050.489826] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.500066] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107103, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.802120] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.850979] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.651s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.851363] env[65121]: DEBUG nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1050.856081] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.479s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.856237] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.858548] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.337s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.858726] env[65121]: DEBUG nova.objects.instance [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1050.900588] env[65121]: INFO nova.scheduler.client.report [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Deleted allocations for instance 27940143-16b5-4263-b23c-354ed8ea8866 [ 1050.931376] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.962906] env[65121]: WARNING neutronclient.v2_0.client [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.000930] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107103, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.263403] env[65121]: DEBUG nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1051.300925] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1051.301502] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1051.301637] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1051.301863] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1051.302104] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1051.302378] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1051.302620] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1051.303018] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1051.303264] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1051.303504] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1051.303792] env[65121]: DEBUG nova.virt.hardware [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1051.305133] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d99836-358e-445d-b2d8-5058b38d88b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.319436] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.320890] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d1b7bd-2013-4460-9684-4ae44e2ebae6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.326238] env[65121]: DEBUG nova.network.neutron [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Port db0cba48-a3d6-4aff-b44f-c1e6672a1daf binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1051.364339] env[65121]: DEBUG nova.compute.utils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1051.369875] env[65121]: DEBUG nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1051.369875] env[65121]: DEBUG nova.network.neutron [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1051.369875] env[65121]: WARNING neutronclient.v2_0.client [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.370145] env[65121]: WARNING neutronclient.v2_0.client [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.370797] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.371220] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.410888] env[65121]: DEBUG oslo_concurrency.lockutils [None req-994aac33-3501-439c-bc54-25135304a2ae tempest-ServersAdminTestJSON-186124780 tempest-ServersAdminTestJSON-186124780-project-member] Lock "27940143-16b5-4263-b23c-354ed8ea8866" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.568s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1051.504262] env[65121]: DEBUG oslo_vmware.api [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107103, 'name': PowerOnVM_Task, 'duration_secs': 0.651101} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.504693] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.504971] env[65121]: INFO nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Took 10.67 seconds to spawn the instance on the hypervisor. [ 1051.505244] env[65121]: DEBUG nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1051.506512] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a3f302-c105-43bd-8886-46cd09bd1af8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.629755] env[65121]: DEBUG nova.policy [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1c51fc43b7f44ce9b023d8cb769e029', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89110618b6ef4ce68b2ca53e7e168139', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1051.801946] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.873379] env[65121]: DEBUG nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1051.879956] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e766d13f-2b4e-4648-ab3c-1a039e006530 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1051.880921] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.496s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1051.882962] env[65121]: INFO nova.compute.claims [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.018632] env[65121]: DEBUG nova.network.neutron [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Successfully updated port: 2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1052.029476] env[65121]: INFO nova.compute.manager [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Took 43.20 seconds to build instance. [ 1052.302762] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.352512] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.352843] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.353046] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.442767] env[65121]: DEBUG nova.network.neutron [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Successfully created port: c73a5f2e-adbf-411c-bec3-31dd92e45737 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1052.522944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.522944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1052.522944] env[65121]: DEBUG nova.network.neutron [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1052.531810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2990aab9-5914-43fe-9af4-9581eb97121a tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "11caa171-0314-4e7d-8502-80932b469da8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.183s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.809209] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.883893] env[65121]: DEBUG nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1052.919816] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1052.919816] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1052.919816] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1052.919816] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1052.919816] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1052.920194] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1052.920194] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.920285] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1052.920695] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1052.920695] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1052.920863] env[65121]: DEBUG nova.virt.hardware [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1052.921849] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3163b3-00e3-4dc6-bfb1-10526545f024 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.933508] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c83b0eb-175f-4c52-9dc3-5357d13d82cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.026821] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.027236] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.233419] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203d6ace-c074-4d8c-9135-ae80ffaa9c8f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.242579] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83169d68-c8f7-4e73-aff8-75d304b9e944 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.282215] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b29918-f4ee-4761-bdab-58887eb74216 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.290227] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056f78cc-bb3e-4ad2-9b71-e25eb1b6e38a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.310121] env[65121]: DEBUG nova.compute.provider_tree [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.315223] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.358762] env[65121]: WARNING neutronclient.v2_0.client [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.365900] env[65121]: DEBUG nova.network.neutron [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1053.413453] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.413861] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1053.413861] env[65121]: DEBUG nova.network.neutron [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1053.431630] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.432060] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.808935] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.817918] env[65121]: DEBUG nova.scheduler.client.report [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1053.918707] env[65121]: WARNING neutronclient.v2_0.client [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.919470] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.919826] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.100187] env[65121]: DEBUG nova.network.neutron [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Successfully updated port: c73a5f2e-adbf-411c-bec3-31dd92e45737 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1054.188731] env[65121]: WARNING neutronclient.v2_0.client [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.189436] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.190856] env[65121]: WARNING openstack [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.200736] env[65121]: DEBUG nova.compute.manager [req-1dbbdaf3-82d2-478e-b97a-5f8d54a46e6d req-c5e5c6ca-c5da-4cee-b6f7-3a0c3bb48315 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Received event network-vif-plugged-2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1054.201306] env[65121]: DEBUG oslo_concurrency.lockutils [req-1dbbdaf3-82d2-478e-b97a-5f8d54a46e6d req-c5e5c6ca-c5da-4cee-b6f7-3a0c3bb48315 service nova] Acquiring lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.201687] env[65121]: DEBUG oslo_concurrency.lockutils [req-1dbbdaf3-82d2-478e-b97a-5f8d54a46e6d req-c5e5c6ca-c5da-4cee-b6f7-3a0c3bb48315 service nova] Lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.201687] env[65121]: DEBUG oslo_concurrency.lockutils [req-1dbbdaf3-82d2-478e-b97a-5f8d54a46e6d req-c5e5c6ca-c5da-4cee-b6f7-3a0c3bb48315 service nova] Lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.201906] env[65121]: DEBUG nova.compute.manager [req-1dbbdaf3-82d2-478e-b97a-5f8d54a46e6d req-c5e5c6ca-c5da-4cee-b6f7-3a0c3bb48315 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] No waiting events found dispatching network-vif-plugged-2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1054.201906] env[65121]: WARNING nova.compute.manager [req-1dbbdaf3-82d2-478e-b97a-5f8d54a46e6d req-c5e5c6ca-c5da-4cee-b6f7-3a0c3bb48315 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Received unexpected event network-vif-plugged-2acd5dbc-a08c-4ebd-922b-284294369a33 for instance with vm_state building and task_state spawning. [ 1054.310023] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.326344] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.326993] env[65121]: DEBUG nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1054.330325] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.179s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.330689] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.333214] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.214s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.333214] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.335485] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.651s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.335610] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.340923] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.421s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.340923] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.340923] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.410s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.341143] env[65121]: DEBUG nova.objects.instance [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lazy-loading 'resources' on Instance uuid 76e4cddb-b046-4a3e-9a91-d3ee847883cf {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.382997] env[65121]: INFO nova.scheduler.client.report [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleted allocations for instance 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3 [ 1054.385115] env[65121]: INFO nova.scheduler.client.report [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Deleted allocations for instance bad70910-58a1-4e35-8d0c-3bd7c3a30abf [ 1054.415360] env[65121]: INFO nova.scheduler.client.report [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Deleted allocations for instance e756fdb2-f901-4e23-92e7-f2753fd5f728 [ 1054.463950] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.464629] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.499731] env[65121]: DEBUG nova.network.neutron [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1054.603451] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "refresh_cache-3d6fc9df-6872-4929-8813-bec1db1ebb16" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.603451] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "refresh_cache-3d6fc9df-6872-4929-8813-bec1db1ebb16" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1054.603451] env[65121]: DEBUG nova.network.neutron [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1054.629750] env[65121]: WARNING neutronclient.v2_0.client [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.630535] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.632020] env[65121]: WARNING openstack [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.811379] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.847272] env[65121]: DEBUG nova.compute.utils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1054.853911] env[65121]: DEBUG nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1054.854146] env[65121]: DEBUG nova.network.neutron [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1054.855101] env[65121]: WARNING neutronclient.v2_0.client [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.855101] env[65121]: WARNING neutronclient.v2_0.client [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.856756] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.857840] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.865084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6f40d1da-0ddf-4a5f-94e7-e11311d22c2f tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.588s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.867467] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 10.514s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.867971] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.868238] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.869106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.871154] env[65121]: INFO nova.compute.manager [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Terminating instance [ 1054.901750] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5fb546bd-869a-412e-b7d9-1e13c0cdc316 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.902985] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d2a71bda-cfb3-4237-bf96-2194b113dd72 tempest-ServerShowV254Test-1220913315 tempest-ServerShowV254Test-1220913315-project-member] Lock "bad70910-58a1-4e35-8d0c-3bd7c3a30abf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.150s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.927045] env[65121]: DEBUG oslo_concurrency.lockutils [None req-89b8a234-6592-419f-a4a8-bea13dfa04b5 tempest-ImagesTestJSON-1950073941 tempest-ImagesTestJSON-1950073941-project-member] Lock "e756fdb2-f901-4e23-92e7-f2753fd5f728" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.499s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.934921] env[65121]: DEBUG nova.network.neutron [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance_info_cache with network_info: [{"id": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "address": "fa:16:3e:d9:2e:bc", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb0cba48-a3", "ovs_interfaceid": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1054.986424] env[65121]: DEBUG nova.compute.manager [req-ec954cbf-4926-4391-9f32-2ba3117acff7 req-ed45c375-99b0-44cb-abcc-fd8096bad298 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Received event network-vif-plugged-c73a5f2e-adbf-411c-bec3-31dd92e45737 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1054.986424] env[65121]: DEBUG oslo_concurrency.lockutils [req-ec954cbf-4926-4391-9f32-2ba3117acff7 req-ed45c375-99b0-44cb-abcc-fd8096bad298 service nova] Acquiring lock "3d6fc9df-6872-4929-8813-bec1db1ebb16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.986424] env[65121]: DEBUG oslo_concurrency.lockutils [req-ec954cbf-4926-4391-9f32-2ba3117acff7 req-ed45c375-99b0-44cb-abcc-fd8096bad298 service nova] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.986424] env[65121]: DEBUG oslo_concurrency.lockutils [req-ec954cbf-4926-4391-9f32-2ba3117acff7 req-ed45c375-99b0-44cb-abcc-fd8096bad298 service nova] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.986424] env[65121]: DEBUG nova.compute.manager [req-ec954cbf-4926-4391-9f32-2ba3117acff7 req-ed45c375-99b0-44cb-abcc-fd8096bad298 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] No waiting events found dispatching network-vif-plugged-c73a5f2e-adbf-411c-bec3-31dd92e45737 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1054.987450] env[65121]: WARNING nova.compute.manager [req-ec954cbf-4926-4391-9f32-2ba3117acff7 req-ed45c375-99b0-44cb-abcc-fd8096bad298 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Received unexpected event network-vif-plugged-c73a5f2e-adbf-411c-bec3-31dd92e45737 for instance with vm_state building and task_state spawning. [ 1054.998306] env[65121]: DEBUG nova.policy [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4853c778ab79419da0e9de23ee3617a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f056059180af48bba4587006efb0b211', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1055.002225] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1055.002567] env[65121]: DEBUG nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Instance network_info: |[{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1055.003019] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:84:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2acd5dbc-a08c-4ebd-922b-284294369a33', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.012823] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1055.016385] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1055.017231] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e4c047e-f059-483e-8cef-598829979d9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.042986] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.042986] env[65121]: value = "task-5107104" [ 1055.042986] env[65121]: _type = "Task" [ 1055.042986] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.054355] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107104, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.085750] env[65121]: DEBUG nova.compute.manager [None req-badb73e1-ec54-4733-9f7c-ec3ce770e3fd tempest-ServerDiagnosticsTest-737919486 tempest-ServerDiagnosticsTest-737919486-project-admin] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1055.087125] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14b0459-b3fc-4e76-8d07-436cf33287a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.097431] env[65121]: INFO nova.compute.manager [None req-badb73e1-ec54-4733-9f7c-ec3ce770e3fd tempest-ServerDiagnosticsTest-737919486 tempest-ServerDiagnosticsTest-737919486-project-admin] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Retrieving diagnostics [ 1055.099715] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926ba778-6ebb-4b5a-8c19-9b8c0349d903 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.139996] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.140470] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.273827] env[65121]: DEBUG nova.network.neutron [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1055.313639] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.331741] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a3159f-6338-4f47-b1b2-cb3b897b238d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.340942] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c51f446-922a-4272-8b17-0446716a856c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.377747] env[65121]: DEBUG nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1055.385035] env[65121]: DEBUG nova.compute.manager [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1055.385035] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.385035] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89f64c4-9b26-4f75-8665-aa062f483289 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.387114] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2f3af24-7398-479b-8e67-75891cd5b903 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.397348] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc3fc98-2eb0-4d07-a7aa-42231487e31c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.405570] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8611fc1-19d7-4603-a3d2-8eb18ce55945 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.430858] env[65121]: DEBUG nova.compute.provider_tree [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.450629] env[65121]: DEBUG oslo_concurrency.lockutils [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1055.458438] env[65121]: WARNING nova.virt.vmwareapi.vmops [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6e969388-3238-404d-a8eb-e7b7318c4c72 could not be found. [ 1055.458438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.458438] env[65121]: INFO nova.compute.manager [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Took 0.07 seconds to destroy the instance on the hypervisor. [ 1055.458438] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1055.458438] env[65121]: DEBUG nova.compute.manager [-] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1055.458438] env[65121]: DEBUG nova.network.neutron [-] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1055.458438] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.459811] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.459811] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.553906] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107104, 'name': CreateVM_Task, 'duration_secs': 0.37603} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.554243] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1055.555305] env[65121]: WARNING neutronclient.v2_0.client [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.555305] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.555399] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1055.555686] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1055.555956] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f26ff288-4ee3-4e5f-92ca-7679d40ace71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.561374] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1055.561374] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525f1ca4-cd32-1e39-9efd-abb15afdbd69" [ 1055.561374] env[65121]: _type = "Task" [ 1055.561374] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.574282] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525f1ca4-cd32-1e39-9efd-abb15afdbd69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.644579] env[65121]: DEBUG nova.network.neutron [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Successfully created port: 65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1055.648654] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.649034] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.705586] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.808420] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.934136] env[65121]: DEBUG nova.scheduler.client.report [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1055.992290] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2649080-7594-4478-9ee9-161053bacaec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.018192] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fbcf5d-dd45-4946-ad6b-3d32a148fc64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.027943] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance '564159fe-6ecd-4276-8d65-cbe25859493a' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1056.075081] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525f1ca4-cd32-1e39-9efd-abb15afdbd69, 'name': SearchDatastore_Task, 'duration_secs': 0.019618} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.075760] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1056.076353] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1056.076353] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.076494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.076620] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1056.076939] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a71126d-72fd-439e-bb80-7d13379b2240 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.087805] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1056.087805] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1056.089131] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e689c432-2983-4db6-b026-9318fc978626 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.099858] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1056.099858] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523e4db8-a807-279c-c05d-454ba1c3d768" [ 1056.099858] env[65121]: _type = "Task" [ 1056.099858] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.108991] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523e4db8-a807-279c-c05d-454ba1c3d768, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.310563] env[65121]: DEBUG oslo_vmware.api [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107102, 'name': ReconfigVM_Task, 'duration_secs': 5.787048} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.310959] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1056.311251] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Reconfigured VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1056.311676] env[65121]: WARNING neutronclient.v2_0.client [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.312058] env[65121]: WARNING neutronclient.v2_0.client [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.312768] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.313175] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.389223] env[65121]: DEBUG nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1056.423336] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1056.423693] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1056.423917] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1056.424148] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1056.424293] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1056.424471] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1056.424667] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.424843] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1056.425028] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1056.425191] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1056.425388] env[65121]: DEBUG nova.virt.hardware [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1056.426387] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f101e99d-907d-4376-9bd8-3908a09c8d1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.437248] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d28562-9274-4729-95ae-a91de353e309 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.440808] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.100s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.473903] env[65121]: INFO nova.scheduler.client.report [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Deleted allocations for instance 76e4cddb-b046-4a3e-9a91-d3ee847883cf [ 1056.475073] env[65121]: DEBUG nova.compute.manager [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1056.536542] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1056.536849] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-459a4602-55cd-4dfa-8464-3530983c9287 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.544799] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1056.544799] env[65121]: value = "task-5107105" [ 1056.544799] env[65121]: _type = "Task" [ 1056.544799] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.555973] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107105, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.610825] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523e4db8-a807-279c-c05d-454ba1c3d768, 'name': SearchDatastore_Task, 'duration_secs': 0.009983} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.611745] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0660dd6b-9237-48d8-8269-a216b2e2cd3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.619360] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1056.619360] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dac641-fcb9-ea6c-3014-150ee6926cec" [ 1056.619360] env[65121]: _type = "Task" [ 1056.619360] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.629467] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dac641-fcb9-ea6c-3014-150ee6926cec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.990708] env[65121]: DEBUG oslo_concurrency.lockutils [None req-30ea7657-de67-4dfb-8922-c9876bc7f7ab tempest-ServersAaction247Test-1467361120 tempest-ServersAaction247Test-1467361120-project-member] Lock "76e4cddb-b046-4a3e-9a91-d3ee847883cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.055s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.007197] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.007197] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.057286] env[65121]: DEBUG oslo_vmware.api [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107105, 'name': PowerOnVM_Task, 'duration_secs': 0.443739} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.057286] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1057.057286] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-986f4e16-036e-498c-aa3f-3e8bd8223bfc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance '564159fe-6ecd-4276-8d65-cbe25859493a' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1057.134626] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dac641-fcb9-ea6c-3014-150ee6926cec, 'name': SearchDatastore_Task, 'duration_secs': 0.022755} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.134626] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.134626] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1057.134626] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ce4de93-04bc-4cc2-8503-ed62c3218f99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.145057] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1057.145057] env[65121]: value = "task-5107106" [ 1057.145057] env[65121]: _type = "Task" [ 1057.145057] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.154047] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.173050] env[65121]: DEBUG nova.network.neutron [-] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.297168] env[65121]: WARNING neutronclient.v2_0.client [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.302059] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.302059] env[65121]: WARNING openstack [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.412515] env[65121]: DEBUG nova.network.neutron [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Successfully updated port: 65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1057.501300] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "03a13029-55b1-4757-80a6-9293471434dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.501541] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "03a13029-55b1-4757-80a6-9293471434dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.510544] env[65121]: INFO nova.compute.claims [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.662338] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107106, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.676789] env[65121]: INFO nova.compute.manager [-] [instance: 6e969388-3238-404d-a8eb-e7b7318c4c72] Took 2.22 seconds to deallocate network for instance. [ 1057.735852] env[65121]: DEBUG nova.network.neutron [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Updating instance_info_cache with network_info: [{"id": "c73a5f2e-adbf-411c-bec3-31dd92e45737", "address": "fa:16:3e:8a:a8:f7", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc73a5f2e-ad", "ovs_interfaceid": "c73a5f2e-adbf-411c-bec3-31dd92e45737", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.920470] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.920709] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1057.920887] env[65121]: DEBUG nova.network.neutron [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1058.004366] env[65121]: DEBUG nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1058.021591] env[65121]: INFO nova.compute.resource_tracker [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating resource usage from migration 06135b5c-3572-417f-99a9-adefc4648d55 [ 1058.162498] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615312} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.166470] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1058.166732] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1058.167916] env[65121]: DEBUG nova.compute.manager [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Received event network-changed-2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1058.168309] env[65121]: DEBUG nova.compute.manager [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Refreshing instance network info cache due to event network-changed-2acd5dbc-a08c-4ebd-922b-284294369a33. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1058.168545] env[65121]: DEBUG oslo_concurrency.lockutils [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Acquiring lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.168684] env[65121]: DEBUG oslo_concurrency.lockutils [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Acquired lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.170153] env[65121]: DEBUG nova.network.neutron [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Refreshing network info cache for port 2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1058.171905] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3550530-6cdc-471b-9331-832ab6f676c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.185564] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1058.185564] env[65121]: value = "task-5107107" [ 1058.185564] env[65121]: _type = "Task" [ 1058.185564] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.193437] env[65121]: DEBUG nova.compute.manager [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Received event network-changed-c73a5f2e-adbf-411c-bec3-31dd92e45737 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1058.193621] env[65121]: DEBUG nova.compute.manager [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Refreshing instance network info cache due to event network-changed-c73a5f2e-adbf-411c-bec3-31dd92e45737. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1058.193879] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Acquiring lock "refresh_cache-3d6fc9df-6872-4929-8813-bec1db1ebb16" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.206512] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.246070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "refresh_cache-3d6fc9df-6872-4929-8813-bec1db1ebb16" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.246070] env[65121]: DEBUG nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Instance network_info: |[{"id": "c73a5f2e-adbf-411c-bec3-31dd92e45737", "address": "fa:16:3e:8a:a8:f7", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc73a5f2e-ad", "ovs_interfaceid": "c73a5f2e-adbf-411c-bec3-31dd92e45737", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1058.246070] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Acquired lock "refresh_cache-3d6fc9df-6872-4929-8813-bec1db1ebb16" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.246070] env[65121]: DEBUG nova.network.neutron [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Refreshing network info cache for port c73a5f2e-adbf-411c-bec3-31dd92e45737 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1058.246070] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:a8:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c73a5f2e-adbf-411c-bec3-31dd92e45737', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.258749] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1058.262583] env[65121]: WARNING neutronclient.v2_0.client [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.263426] env[65121]: WARNING openstack [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.263556] env[65121]: WARNING openstack [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.275155] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.276459] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d95d306b-1084-443a-a8fb-496b7fe0128a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.301801] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.301801] env[65121]: value = "task-5107108" [ 1058.301801] env[65121]: _type = "Task" [ 1058.301801] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.314726] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107108, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.425316] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.425853] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.494898] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00754309-c81c-491a-8a3a-ad21256f5a23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.506535] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220827cd-6c60-4c8d-b206-e694ef6e87eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.544330] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.546407] env[65121]: DEBUG nova.network.neutron [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1058.548355] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0b5339-f8a9-4460-8e66-6eeaeee3fd38 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.556649] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd2493c-cb94-48f9-89f9-3321d7d40e6c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.583205] env[65121]: DEBUG nova.compute.provider_tree [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.609125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.609125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.609645] env[65121]: DEBUG nova.network.neutron [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1058.614450] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "11caa171-0314-4e7d-8502-80932b469da8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.614450] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "11caa171-0314-4e7d-8502-80932b469da8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.614706] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "11caa171-0314-4e7d-8502-80932b469da8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.614945] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "11caa171-0314-4e7d-8502-80932b469da8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.615668] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "11caa171-0314-4e7d-8502-80932b469da8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.619415] env[65121]: INFO nova.compute.manager [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Terminating instance [ 1058.663220] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.663781] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.672154] env[65121]: WARNING neutronclient.v2_0.client [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.672737] env[65121]: WARNING openstack [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.673161] env[65121]: WARNING openstack [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.701993] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101239} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.703316] env[65121]: WARNING openstack [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.703683] env[65121]: WARNING openstack [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.710084] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1058.711514] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0015a990-aee8-48cb-9dbb-d27790bfe7d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.739047] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1058.739673] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76c1a58d-4e5f-4d45-94eb-789bd711f157 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "6e969388-3238-404d-a8eb-e7b7318c4c72" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.873s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.740633] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-663ff44e-3413-4df5-aeb2-7008cda8745a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.769312] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1058.769312] env[65121]: value = "task-5107109" [ 1058.769312] env[65121]: _type = "Task" [ 1058.769312] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.782177] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107109, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.812199] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107108, 'name': CreateVM_Task, 'duration_secs': 0.367765} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.812456] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.813029] env[65121]: WARNING neutronclient.v2_0.client [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.813357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.813576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.813810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1058.814087] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3109b4c-620e-4f4a-a576-975b0531426b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.822436] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1058.822436] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527c4e64-be07-fc7c-ae60-5bbcce3ef511" [ 1058.822436] env[65121]: _type = "Task" [ 1058.822436] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.835045] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527c4e64-be07-fc7c-ae60-5bbcce3ef511, 'name': SearchDatastore_Task, 'duration_secs': 0.010555} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.835045] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.836469] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.836469] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.836469] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.836469] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.836716] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2abfc5b-ce79-4837-b717-022cc6636b4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.847618] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.847618] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.847618] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-950c89e9-f630-4f2d-b718-87ad3889caf1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.854014] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1058.854014] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529b1f56-0092-0a3e-e8a1-5a68089ab4fe" [ 1058.854014] env[65121]: _type = "Task" [ 1058.854014] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.864209] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529b1f56-0092-0a3e-e8a1-5a68089ab4fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.906464] env[65121]: WARNING neutronclient.v2_0.client [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.907872] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.907872] env[65121]: WARNING openstack [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.935532] env[65121]: WARNING neutronclient.v2_0.client [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.936484] env[65121]: WARNING openstack [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.937101] env[65121]: WARNING openstack [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.032520] env[65121]: WARNING openstack [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.032907] env[65121]: WARNING openstack [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.089317] env[65121]: DEBUG nova.scheduler.client.report [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1059.100494] env[65121]: DEBUG nova.network.neutron [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Updated VIF entry in instance network info cache for port c73a5f2e-adbf-411c-bec3-31dd92e45737. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1059.100632] env[65121]: DEBUG nova.network.neutron [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Updating instance_info_cache with network_info: [{"id": "c73a5f2e-adbf-411c-bec3-31dd92e45737", "address": "fa:16:3e:8a:a8:f7", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc73a5f2e-ad", "ovs_interfaceid": "c73a5f2e-adbf-411c-bec3-31dd92e45737", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.104762] env[65121]: DEBUG nova.network.neutron [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Updating instance_info_cache with network_info: [{"id": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "address": "fa:16:3e:34:30:0f", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c052c9-c5", "ovs_interfaceid": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.114096] env[65121]: WARNING neutronclient.v2_0.client [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.114096] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.114096] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.128461] env[65121]: DEBUG nova.compute.manager [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1059.128461] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.128461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd37601-3435-452a-bd55-f4fb9961e801 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.144149] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.145438] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16f941bb-ca9e-48d1-861b-16aea7134f5c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.153312] env[65121]: DEBUG oslo_vmware.api [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1059.153312] env[65121]: value = "task-5107110" [ 1059.153312] env[65121]: _type = "Task" [ 1059.153312] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.163198] env[65121]: DEBUG oslo_vmware.api [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107110, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.208968] env[65121]: WARNING neutronclient.v2_0.client [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.211035] env[65121]: WARNING openstack [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.211035] env[65121]: WARNING openstack [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.284214] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107109, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.340617] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.341120] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.350164] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.350692] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "23d6f632-8918-46af-b239-08a9615dfbec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.351019] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "23d6f632-8918-46af-b239-08a9615dfbec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.351294] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.351547] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "23d6f632-8918-46af-b239-08a9615dfbec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1059.354933] env[65121]: INFO nova.compute.manager [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Terminating instance [ 1059.372271] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529b1f56-0092-0a3e-e8a1-5a68089ab4fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009803} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.373876] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f030632-d56f-4b76-becd-07c536cd3fe3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.382067] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1059.382067] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f83c91-1246-9374-75c9-f0911afea2fb" [ 1059.382067] env[65121]: _type = "Task" [ 1059.382067] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.395521] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f83c91-1246-9374-75c9-f0911afea2fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.412025] env[65121]: DEBUG nova.network.neutron [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updated VIF entry in instance network info cache for port 2acd5dbc-a08c-4ebd-922b-284294369a33. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1059.412025] env[65121]: DEBUG nova.network.neutron [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.520051] env[65121]: WARNING neutronclient.v2_0.client [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.520346] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.520588] env[65121]: WARNING openstack [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.597161] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.591s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1059.597355] env[65121]: INFO nova.compute.manager [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Migrating [ 1059.604839] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.061s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.606365] env[65121]: INFO nova.compute.claims [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.611339] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ef99be6-83ae-4afd-bb86-d64c97f8811e req-bb3f3baf-5368-4c07-85bc-cf4b39a25883 service nova] Releasing lock "refresh_cache-3d6fc9df-6872-4929-8813-bec1db1ebb16" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.611830] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.612186] env[65121]: DEBUG nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Instance network_info: |[{"id": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "address": "fa:16:3e:34:30:0f", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c052c9-c5", "ovs_interfaceid": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1059.617417] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:30:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65c052c9-c53e-40b1-a26b-6b8e80093cee', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.628584] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1059.634417] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.635699] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5924a2fc-6d01-4142-84eb-c7778f903eb6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.669654] env[65121]: DEBUG oslo_vmware.api [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107110, 'name': PowerOffVM_Task, 'duration_secs': 0.2298} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.671281] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.671613] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.672261] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.672261] env[65121]: value = "task-5107111" [ 1059.672261] env[65121]: _type = "Task" [ 1059.672261] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.672572] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abd72a08-fd83-4a4d-812b-817ce554042f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.685568] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107111, 'name': CreateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.708086] env[65121]: INFO nova.network.neutron [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Port 15f422aa-08f5-4473-b1ac-92eb03870e58 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1059.708316] env[65121]: INFO nova.network.neutron [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Port 065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1059.708675] env[65121]: DEBUG nova.network.neutron [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [{"id": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "address": "fa:16:3e:24:b9:43", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf696eb6b-11", "ovs_interfaceid": "f696eb6b-11cc-4a52-9742-e5de719ed10b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.782785] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107109, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.847347] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.847659] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.847850] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Deleting the datastore file [datastore1] 11caa171-0314-4e7d-8502-80932b469da8 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.848194] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4926a184-f636-4681-8b05-a90d0288ccf7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.858217] env[65121]: DEBUG oslo_vmware.api [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for the task: (returnval){ [ 1059.858217] env[65121]: value = "task-5107113" [ 1059.858217] env[65121]: _type = "Task" [ 1059.858217] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.863664] env[65121]: DEBUG nova.compute.manager [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1059.864020] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.869444] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24990972-39a2-42e2-a20d-593713e72646 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.874040] env[65121]: DEBUG oslo_vmware.api [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.879852] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.880188] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f6208ee-a6ba-4cee-9a3c-8f6d11faafc3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.887828] env[65121]: DEBUG oslo_vmware.api [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1059.887828] env[65121]: value = "task-5107114" [ 1059.887828] env[65121]: _type = "Task" [ 1059.887828] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.897168] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f83c91-1246-9374-75c9-f0911afea2fb, 'name': SearchDatastore_Task, 'duration_secs': 0.022177} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.898137] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.898577] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 3d6fc9df-6872-4929-8813-bec1db1ebb16/3d6fc9df-6872-4929-8813-bec1db1ebb16.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.898936] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f27fe2ec-9dc1-418a-9414-e30aa34d959e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.905611] env[65121]: DEBUG oslo_vmware.api [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.911778] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1059.911778] env[65121]: value = "task-5107115" [ 1059.911778] env[65121]: _type = "Task" [ 1059.911778] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.921498] env[65121]: DEBUG oslo_concurrency.lockutils [req-3f9eceb4-aa6b-476e-a34e-0c43e53017a1 req-f80a6b06-a731-480e-90c3-ff0ed5aa8bf0 service nova] Releasing lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.921949] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107115, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.117222] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.117392] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.117560] env[65121]: DEBUG nova.network.neutron [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1060.190057] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107111, 'name': CreateVM_Task, 'duration_secs': 0.402663} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.190057] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.190240] env[65121]: WARNING neutronclient.v2_0.client [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.190521] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.190674] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.191019] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1060.191315] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-551e7e0b-0121-42f0-a8ab-6f3c2c8747c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.202189] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1060.202189] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f14588-908b-5074-6c82-fa082641359c" [ 1060.202189] env[65121]: _type = "Task" [ 1060.202189] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.211378] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-23d6f632-8918-46af-b239-08a9615dfbec" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1060.220511] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f14588-908b-5074-6c82-fa082641359c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.285624] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107109, 'name': ReconfigVM_Task, 'duration_secs': 1.133983} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.286041] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.286725] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d56ff138-b3f0-4833-af32-92de1ce9dc2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.296855] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1060.296855] env[65121]: value = "task-5107116" [ 1060.296855] env[65121]: _type = "Task" [ 1060.296855] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.308526] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107116, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.312528] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "564159fe-6ecd-4276-8d65-cbe25859493a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.314119] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.314119] env[65121]: DEBUG nova.compute.manager [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Going to confirm migration 4 {{(pid=65121) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1060.369798] env[65121]: DEBUG oslo_vmware.api [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Task: {'id': task-5107113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238439} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.370135] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.370358] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.370546] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.370711] env[65121]: INFO nova.compute.manager [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Took 1.25 seconds to destroy the instance on the hypervisor. [ 1060.370966] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1060.371264] env[65121]: DEBUG nova.compute.manager [-] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1060.371331] env[65121]: DEBUG nova.network.neutron [-] [instance: 11caa171-0314-4e7d-8502-80932b469da8] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1060.371635] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.372322] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.372684] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.400911] env[65121]: DEBUG oslo_vmware.api [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107114, 'name': PowerOffVM_Task, 'duration_secs': 0.249199} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.401292] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1060.401470] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1060.401850] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afc37bbd-cb6d-4e91-9c56-2dab369f5e1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.423361] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107115, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.473143] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.531200] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1060.531608] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1060.531838] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleting the datastore file [datastore1] 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.532186] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba5cb7bc-93c9-4dd7-b015-c2c2f7b4ac10 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.542647] env[65121]: DEBUG oslo_vmware.api [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1060.542647] env[65121]: value = "task-5107118" [ 1060.542647] env[65121]: _type = "Task" [ 1060.542647] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.556364] env[65121]: DEBUG oslo_vmware.api [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107118, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.622187] env[65121]: WARNING neutronclient.v2_0.client [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.623065] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.623422] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.718874] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f14588-908b-5074-6c82-fa082641359c, 'name': SearchDatastore_Task, 'duration_secs': 0.061866} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.719457] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1060.719704] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.719965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.720125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.720305] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.720591] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-408975ba-dddc-454f-80e7-1a28e7878111 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.723461] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f43ba648-025e-4d71-a189-7a94a7e2c085 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-23d6f632-8918-46af-b239-08a9615dfbec-15f422aa-08f5-4473-b1ac-92eb03870e58" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.041s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.748242] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.748242] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.748396] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbb121eb-a313-4638-9f4f-aaff03136c98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.757365] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1060.757365] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c2142f-5dda-f8d6-1f3e-1bca3cec8150" [ 1060.757365] env[65121]: _type = "Task" [ 1060.757365] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.774754] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c2142f-5dda-f8d6-1f3e-1bca3cec8150, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.812899] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107116, 'name': Rename_Task, 'duration_secs': 0.247924} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.815774] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.818888] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9c21c4e-2892-4aad-8cdc-9da2e9b7b1ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.821718] env[65121]: WARNING neutronclient.v2_0.client [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.833667] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1060.833667] env[65121]: value = "task-5107119" [ 1060.833667] env[65121]: _type = "Task" [ 1060.833667] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.845637] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107119, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.848852] env[65121]: DEBUG nova.compute.manager [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Received event network-vif-plugged-65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1060.848852] env[65121]: DEBUG oslo_concurrency.lockutils [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Acquiring lock "fc3109cb-72bb-4695-82c3-a323e619919d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.848852] env[65121]: DEBUG oslo_concurrency.lockutils [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Lock "fc3109cb-72bb-4695-82c3-a323e619919d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.848852] env[65121]: DEBUG oslo_concurrency.lockutils [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Lock "fc3109cb-72bb-4695-82c3-a323e619919d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.849090] env[65121]: DEBUG nova.compute.manager [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] No waiting events found dispatching network-vif-plugged-65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1060.849239] env[65121]: WARNING nova.compute.manager [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Received unexpected event network-vif-plugged-65c052c9-c53e-40b1-a26b-6b8e80093cee for instance with vm_state building and task_state spawning. [ 1060.849359] env[65121]: DEBUG nova.compute.manager [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Received event network-changed-65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1060.849495] env[65121]: DEBUG nova.compute.manager [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Refreshing instance network info cache due to event network-changed-65c052c9-c53e-40b1-a26b-6b8e80093cee. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1060.849658] env[65121]: DEBUG oslo_concurrency.lockutils [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Acquiring lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.849777] env[65121]: DEBUG oslo_concurrency.lockutils [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Acquired lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.849916] env[65121]: DEBUG nova.network.neutron [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Refreshing network info cache for port 65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1060.891960] env[65121]: DEBUG nova.compute.manager [req-6c388a34-e681-4a31-a422-eb623fff4090 req-aa82429a-79eb-4be5-bb97-669faeee37b3 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-vif-deleted-15f422aa-08f5-4473-b1ac-92eb03870e58 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1060.923688] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107115, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596082} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.927735] env[65121]: WARNING neutronclient.v2_0.client [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.928133] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.928318] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.928442] env[65121]: DEBUG nova.network.neutron [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1060.928611] env[65121]: DEBUG nova.objects.instance [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lazy-loading 'info_cache' on Instance uuid 564159fe-6ecd-4276-8d65-cbe25859493a {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.930732] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 3d6fc9df-6872-4929-8813-bec1db1ebb16/3d6fc9df-6872-4929-8813-bec1db1ebb16.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1060.930994] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.931940] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a80bf7f3-c917-4243-8b9e-407382851ec9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.939767] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1060.939767] env[65121]: value = "task-5107120" [ 1060.939767] env[65121]: _type = "Task" [ 1060.939767] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.953300] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107120, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.958421] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.958812] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.002513] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861a2099-791f-4ee3-8f4e-bd1879aa1b22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.012201] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12de2fca-c7f1-4144-8835-290dad53d588 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.056297] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a369f248-1dea-4af7-a207-4acdea41ad45 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.065642] env[65121]: DEBUG oslo_vmware.api [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107118, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210916} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.067919] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.069426] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.069426] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.069426] env[65121]: INFO nova.compute.manager [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1061.069426] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1061.069757] env[65121]: DEBUG nova.compute.manager [-] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1061.069757] env[65121]: DEBUG nova.network.neutron [-] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1061.070023] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.070749] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.071222] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.080178] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6973372-c87f-4580-88a2-dc91273fb6af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.100889] env[65121]: DEBUG nova.compute.provider_tree [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.190332] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.192556] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.193572] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.240204] env[65121]: WARNING neutronclient.v2_0.client [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.240953] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.241924] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.274183] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c2142f-5dda-f8d6-1f3e-1bca3cec8150, 'name': SearchDatastore_Task, 'duration_secs': 0.019377} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.276238] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3af534b-d748-46fa-86bd-69bb96b3fa11 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.283026] env[65121]: DEBUG nova.network.neutron [-] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1061.284344] env[65121]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 could not be found.", "detail": ""}} {{(pid=65121) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:265}} [ 1061.284667] env[65121]: DEBUG nova.network.neutron [-] Unable to show port 065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 as it no longer exists. {{(pid=65121) _unbind_ports /opt/stack/nova/nova/network/neutron.py:700}} [ 1061.295570] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1061.295570] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dfdde1-5ba5-621a-0b09-0badce09ece2" [ 1061.295570] env[65121]: _type = "Task" [ 1061.295570] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.306876] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dfdde1-5ba5-621a-0b09-0badce09ece2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.347622] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107119, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.354144] env[65121]: WARNING neutronclient.v2_0.client [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.354821] env[65121]: WARNING openstack [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.355809] env[65121]: WARNING openstack [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.450492] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107120, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082335} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.450764] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.451609] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782fc92e-8769-432c-9e46-b8efa86ffc67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.387591] env[65121]: DEBUG nova.scheduler.client.report [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1062.391221] env[65121]: DEBUG nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1062.394167] env[65121]: INFO nova.compute.manager [-] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Took 2.02 seconds to deallocate network for instance. [ 1062.395043] env[65121]: WARNING neutronclient.v2_0.client [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.395806] env[65121]: WARNING openstack [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.396187] env[65121]: WARNING openstack [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.413582] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 3d6fc9df-6872-4929-8813-bec1db1ebb16/3d6fc9df-6872-4929-8813-bec1db1ebb16.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.419869] env[65121]: DEBUG nova.network.neutron [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1062.431896] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7c307b6-64b1-4399-8258-163b43f9d9f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.447412] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.459838] env[65121]: DEBUG oslo_vmware.api [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107119, 'name': PowerOnVM_Task, 'duration_secs': 1.101781} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.464286] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1062.464773] env[65121]: INFO nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Took 11.20 seconds to spawn the instance on the hypervisor. [ 1062.464773] env[65121]: DEBUG nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1062.465106] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1062.465106] env[65121]: value = "task-5107121" [ 1062.465106] env[65121]: _type = "Task" [ 1062.465106] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.465452] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dfdde1-5ba5-621a-0b09-0badce09ece2, 'name': SearchDatastore_Task, 'duration_secs': 0.015969} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.466035] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af5b0ea-bb2b-4517-98b6-5635c1288fc4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.471207] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.471207] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] fc3109cb-72bb-4695-82c3-a323e619919d/fc3109cb-72bb-4695-82c3-a323e619919d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.476582] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-305a6f64-71ca-49f1-b522-3668295bd4e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.488909] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107121, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.492238] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1062.492238] env[65121]: value = "task-5107122" [ 1062.492238] env[65121]: _type = "Task" [ 1062.492238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.506147] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107122, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.601641] env[65121]: WARNING openstack [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.602157] env[65121]: WARNING openstack [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.677684] env[65121]: WARNING openstack [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.677684] env[65121]: WARNING openstack [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.832278] env[65121]: WARNING neutronclient.v2_0.client [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.832977] env[65121]: WARNING openstack [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.833340] env[65121]: WARNING openstack [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.919863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.315s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.920467] env[65121]: DEBUG nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1062.954488] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.955076] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.958560] env[65121]: INFO nova.compute.claims [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.968319] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.968319] env[65121]: DEBUG nova.network.neutron [-] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1062.980821] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107121, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.983399] env[65121]: WARNING neutronclient.v2_0.client [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.984140] env[65121]: WARNING openstack [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.984859] env[65121]: WARNING openstack [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1063.014026] env[65121]: INFO nova.compute.manager [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Took 48.04 seconds to build instance. [ 1063.016941] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107122, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.201441] env[65121]: DEBUG nova.network.neutron [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Updated VIF entry in instance network info cache for port 65c052c9-c53e-40b1-a26b-6b8e80093cee. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1063.201633] env[65121]: DEBUG nova.network.neutron [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Updating instance_info_cache with network_info: [{"id": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "address": "fa:16:3e:34:30:0f", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c052c9-c5", "ovs_interfaceid": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.241144] env[65121]: DEBUG nova.network.neutron [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance_info_cache with network_info: [{"id": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "address": "fa:16:3e:d9:2e:bc", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb0cba48-a3", "ovs_interfaceid": "db0cba48-a3d6-4aff-b44f-c1e6672a1daf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.353273] env[65121]: DEBUG nova.compute.manager [req-dede58cc-2ef3-499b-ae42-5db3f88e85cb req-034fd514-e7a7-46e8-a8b8-7ae86a051a50 service nova] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Received event network-vif-deleted-59929561-be86-4157-85cc-e3e04892acbc {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1063.353273] env[65121]: DEBUG nova.compute.manager [req-dede58cc-2ef3-499b-ae42-5db3f88e85cb req-034fd514-e7a7-46e8-a8b8-7ae86a051a50 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-vif-deleted-f696eb6b-11cc-4a52-9742-e5de719ed10b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1063.428624] env[65121]: DEBUG nova.compute.utils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1063.432817] env[65121]: DEBUG nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1063.432817] env[65121]: DEBUG nova.network.neutron [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1063.432817] env[65121]: WARNING neutronclient.v2_0.client [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1063.432817] env[65121]: WARNING neutronclient.v2_0.client [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1063.433328] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1063.433704] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1063.477807] env[65121]: INFO nova.compute.manager [-] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Took 2.41 seconds to deallocate network for instance. [ 1063.496775] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107121, 'name': ReconfigVM_Task, 'duration_secs': 0.644262} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.496775] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 3d6fc9df-6872-4929-8813-bec1db1ebb16/3d6fc9df-6872-4929-8813-bec1db1ebb16.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.496775] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62ace46a-85f9-4dff-8950-ccbc49cfd89e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.507631] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107122, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534294} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.508432] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] fc3109cb-72bb-4695-82c3-a323e619919d/fc3109cb-72bb-4695-82c3-a323e619919d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.508656] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.508968] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1063.508968] env[65121]: value = "task-5107123" [ 1063.508968] env[65121]: _type = "Task" [ 1063.508968] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.509177] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ccc2e3d-8d5e-4a4a-9aa6-85845233a1b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.523835] env[65121]: DEBUG oslo_concurrency.lockutils [None req-250a0ace-1b29-4d1b-95d3-d168b7973968 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.893s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.523835] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107123, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.525071] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1063.525071] env[65121]: value = "task-5107124" [ 1063.525071] env[65121]: _type = "Task" [ 1063.525071] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.535299] env[65121]: DEBUG nova.policy [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f9ed936951a4078b56e0b188cd8794f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a98eae6244314b09865c29eff2ee56a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1063.542459] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107124, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.704896] env[65121]: DEBUG oslo_concurrency.lockutils [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] Releasing lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.706343] env[65121]: DEBUG nova.compute.manager [req-f1e204a6-190b-4da1-bb28-5cc04d5c929d req-257bfab4-178f-496a-a33a-91d61150da87 service nova] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Received event network-vif-deleted-065401c7-8dd5-4be7-a65a-d1c11a0d2ba6 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1063.744652] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-564159fe-6ecd-4276-8d65-cbe25859493a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.745787] env[65121]: DEBUG nova.objects.instance [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lazy-loading 'migration_context' on Instance uuid 564159fe-6ecd-4276-8d65-cbe25859493a {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.765666] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1063.765971] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.914413] env[65121]: DEBUG nova.network.neutron [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Successfully created port: 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1063.933495] env[65121]: DEBUG nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1063.984559] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956d5b6f-e1c8-4188-b365-7992807da7ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.989894] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.012142] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance 'c4096314-270a-4270-9e1d-5ace8ddbd286' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1064.028503] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107123, 'name': Rename_Task, 'duration_secs': 0.308013} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.032576] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.033072] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a767fb77-fa64-4a44-a929-182d32f5de61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.049305] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107124, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069573} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.050706] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.051124] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1064.051124] env[65121]: value = "task-5107125" [ 1064.051124] env[65121]: _type = "Task" [ 1064.051124] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.055992] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d662604f-25c5-4e50-87e0-bc17b46dc6fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.068953] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107125, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.089925] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] fc3109cb-72bb-4695-82c3-a323e619919d/fc3109cb-72bb-4695-82c3-a323e619919d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.092595] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-037c6096-1157-43c4-80c7-5994a62329bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.115692] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1064.115692] env[65121]: value = "task-5107126" [ 1064.115692] env[65121]: _type = "Task" [ 1064.115692] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.126487] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107126, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.248747] env[65121]: DEBUG nova.objects.base [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Object Instance<564159fe-6ecd-4276-8d65-cbe25859493a> lazy-loaded attributes: info_cache,migration_context {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1064.249763] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340277f2-b7a9-463d-ab60-f6bdd21533e6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.279981] env[65121]: DEBUG nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1064.287269] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf66774f-fede-4948-afae-3af561878b1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.295524] env[65121]: DEBUG oslo_vmware.api [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1064.295524] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52330d4e-5b60-f079-f9e0-ee7edd2a98b1" [ 1064.295524] env[65121]: _type = "Task" [ 1064.295524] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.308417] env[65121]: DEBUG oslo_vmware.api [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52330d4e-5b60-f079-f9e0-ee7edd2a98b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.440508] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a11527-1699-4ae0-8c42-cc161c6df6b0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.454613] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491c909a-0b22-4a3a-a687-d32d98e65688 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.494138] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c610768-57c6-43b2-8808-8d8911b4ec5a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.505044] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d96097-c250-4117-8d6d-1682fe44897e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.522161] env[65121]: DEBUG nova.compute.provider_tree [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1064.526206] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1064.526394] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-caeb8bb6-c02d-4847-8d31-dc9dc990632a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.536239] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1064.536239] env[65121]: value = "task-5107127" [ 1064.536239] env[65121]: _type = "Task" [ 1064.536239] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.549614] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.571438] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107125, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.629115] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107126, 'name': ReconfigVM_Task, 'duration_secs': 0.379666} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.629500] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Reconfigured VM instance instance-00000061 to attach disk [datastore1] fc3109cb-72bb-4695-82c3-a323e619919d/fc3109cb-72bb-4695-82c3-a323e619919d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.630389] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f10d119-6456-451c-b23d-3d5e14937938 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.640173] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1064.640173] env[65121]: value = "task-5107128" [ 1064.640173] env[65121]: _type = "Task" [ 1064.640173] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.651325] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107128, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.808735] env[65121]: DEBUG oslo_vmware.api [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52330d4e-5b60-f079-f9e0-ee7edd2a98b1, 'name': SearchDatastore_Task, 'duration_secs': 0.01076} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.809272] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.825608] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.953653] env[65121]: DEBUG nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1064.987153] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1064.987498] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1064.987695] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1064.988014] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1064.988266] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1064.989026] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1064.989026] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.989026] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1064.989026] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1064.989307] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1064.989474] env[65121]: DEBUG nova.virt.hardware [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1064.990461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b2f635-a513-4957-a3e0-47e34d5366a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.999826] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e90cdb-8aca-4d86-98a7-d5daa50a855a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.048824] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107127, 'name': PowerOffVM_Task, 'duration_secs': 0.312877} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.048824] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.048824] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance 'c4096314-270a-4270-9e1d-5ace8ddbd286' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1065.054996] env[65121]: ERROR nova.scheduler.client.report [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [req-86bfdde8-73c9-4b15-9790-07d91fafe55e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-86bfdde8-73c9-4b15-9790-07d91fafe55e"}]} [ 1065.069053] env[65121]: DEBUG oslo_vmware.api [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107125, 'name': PowerOnVM_Task, 'duration_secs': 0.631146} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.069588] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.069588] env[65121]: INFO nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Took 12.19 seconds to spawn the instance on the hypervisor. [ 1065.069838] env[65121]: DEBUG nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1065.070562] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec44b2da-d8a4-4883-bcf5-dcdcbe610adc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.083149] env[65121]: DEBUG nova.scheduler.client.report [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1065.110386] env[65121]: DEBUG nova.scheduler.client.report [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1065.110386] env[65121]: DEBUG nova.compute.provider_tree [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1065.126092] env[65121]: DEBUG nova.scheduler.client.report [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1065.149868] env[65121]: DEBUG nova.scheduler.client.report [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1065.158261] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107128, 'name': Rename_Task, 'duration_secs': 0.218551} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.158581] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.158849] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b968c15e-8542-4a76-83a7-917703282936 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.168374] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1065.168374] env[65121]: value = "task-5107129" [ 1065.168374] env[65121]: _type = "Task" [ 1065.168374] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.182732] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.381678] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "ec463cad-4c80-4636-bc7a-9ec298a07d96" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.381678] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.381893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "ec463cad-4c80-4636-bc7a-9ec298a07d96-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.382044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.382525] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.385098] env[65121]: INFO nova.compute.manager [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Terminating instance [ 1065.560583] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1065.560964] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1065.561027] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1065.561164] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1065.561305] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1065.561440] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1065.561639] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1065.561782] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1065.561938] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1065.562425] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1065.562425] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1065.572881] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48e61899-93fb-4ddf-a36d-7a5ff9703267 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.592596] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1065.592596] env[65121]: value = "task-5107130" [ 1065.592596] env[65121]: _type = "Task" [ 1065.592596] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.598048] env[65121]: INFO nova.compute.manager [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Took 49.16 seconds to build instance. [ 1065.601283] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d995a0-ed25-4669-a501-54bdc8624eec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.608845] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107130, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.616125] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ff248f-c825-4b5b-891e-88a3047c96a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.660838] env[65121]: DEBUG nova.network.neutron [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Successfully updated port: 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1065.662607] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442deebc-bfe0-402e-951c-02663d394bb3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.676103] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd3827a-557f-424b-a495-1823f511e9dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.688452] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.697467] env[65121]: DEBUG nova.compute.provider_tree [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1065.895025] env[65121]: DEBUG nova.compute.manager [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1065.895025] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1065.895025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29035d59-ce3d-4ff8-80e1-71a232760a90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.906158] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.906996] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fb2db1d-f90e-4dca-ba9a-df896882d2db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.918756] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1065.918756] env[65121]: value = "task-5107131" [ 1065.918756] env[65121]: _type = "Task" [ 1065.918756] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.930359] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.978825] env[65121]: DEBUG nova.compute.manager [req-2941d9e7-9170-422e-a81d-af3560fa3ca5 req-00feec86-4d47-435f-9b96-71f1463d15e1 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received event network-vif-plugged-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1065.979084] env[65121]: DEBUG oslo_concurrency.lockutils [req-2941d9e7-9170-422e-a81d-af3560fa3ca5 req-00feec86-4d47-435f-9b96-71f1463d15e1 service nova] Acquiring lock "03a13029-55b1-4757-80a6-9293471434dc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.979307] env[65121]: DEBUG oslo_concurrency.lockutils [req-2941d9e7-9170-422e-a81d-af3560fa3ca5 req-00feec86-4d47-435f-9b96-71f1463d15e1 service nova] Lock "03a13029-55b1-4757-80a6-9293471434dc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.979458] env[65121]: DEBUG oslo_concurrency.lockutils [req-2941d9e7-9170-422e-a81d-af3560fa3ca5 req-00feec86-4d47-435f-9b96-71f1463d15e1 service nova] Lock "03a13029-55b1-4757-80a6-9293471434dc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.979614] env[65121]: DEBUG nova.compute.manager [req-2941d9e7-9170-422e-a81d-af3560fa3ca5 req-00feec86-4d47-435f-9b96-71f1463d15e1 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] No waiting events found dispatching network-vif-plugged-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1065.979792] env[65121]: WARNING nova.compute.manager [req-2941d9e7-9170-422e-a81d-af3560fa3ca5 req-00feec86-4d47-435f-9b96-71f1463d15e1 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received unexpected event network-vif-plugged-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 for instance with vm_state building and task_state spawning. [ 1066.013940] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.013940] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.014184] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.014364] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.014524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.017540] env[65121]: INFO nova.compute.manager [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Terminating instance [ 1066.115868] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6816d0ed-9fcf-477a-87b7-4c7a28cc7ed3 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.779s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.116520] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107130, 'name': ReconfigVM_Task, 'duration_secs': 0.382481} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.117136] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance 'c4096314-270a-4270-9e1d-5ace8ddbd286' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1066.166759] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.167346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquired lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.167694] env[65121]: DEBUG nova.network.neutron [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1066.186426] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.317209] env[65121]: DEBUG nova.scheduler.client.report [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 132 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1066.317482] env[65121]: DEBUG nova.compute.provider_tree [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 132 to 133 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1066.317656] env[65121]: DEBUG nova.compute.provider_tree [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1066.432444] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107131, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.520090] env[65121]: DEBUG nova.compute.manager [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Received event network-changed-2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1066.520090] env[65121]: DEBUG nova.compute.manager [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Refreshing instance network info cache due to event network-changed-2acd5dbc-a08c-4ebd-922b-284294369a33. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1066.520090] env[65121]: DEBUG oslo_concurrency.lockutils [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Acquiring lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.520245] env[65121]: DEBUG oslo_concurrency.lockutils [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Acquired lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.520483] env[65121]: DEBUG nova.network.neutron [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Refreshing network info cache for port 2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1066.522448] env[65121]: DEBUG nova.compute.manager [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1066.522691] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.523977] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2058992d-f8a9-4fc1-a0be-0a0edd7aaa61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.532972] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.533389] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56ee6547-1e68-4fe3-a566-6b644cc49c83 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.541980] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1066.541980] env[65121]: value = "task-5107132" [ 1066.541980] env[65121]: _type = "Task" [ 1066.541980] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.555398] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.632948] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1066.633322] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1066.633422] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1066.633690] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1066.633847] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1066.633988] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1066.634286] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.634484] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1066.634690] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1066.634907] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1066.635163] env[65121]: DEBUG nova.virt.hardware [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1066.642132] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Reconfiguring VM instance instance-0000000d to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1066.642132] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4d43f8e-626d-4165-bfd7-d013fb0210f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.666989] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1066.666989] env[65121]: value = "task-5107133" [ 1066.666989] env[65121]: _type = "Task" [ 1066.666989] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.671723] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1066.672538] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1066.689923] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.696669] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.718396] env[65121]: DEBUG nova.network.neutron [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1066.748538] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1066.748980] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1066.804561] env[65121]: WARNING neutronclient.v2_0.client [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1066.805246] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1066.805636] env[65121]: WARNING openstack [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1066.826042] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.870s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.826042] env[65121]: DEBUG nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1066.834989] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.867s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.835094] env[65121]: DEBUG nova.objects.instance [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lazy-loading 'resources' on Instance uuid 11caa171-0314-4e7d-8502-80932b469da8 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.911754] env[65121]: DEBUG nova.network.neutron [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1066.930875] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107131, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.975095] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "39ca7a89-c262-4169-91b1-92ea45744518" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.975740] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "39ca7a89-c262-4169-91b1-92ea45744518" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.028336] env[65121]: WARNING neutronclient.v2_0.client [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.029023] env[65121]: WARNING openstack [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.029434] env[65121]: WARNING openstack [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.058657] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107132, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.139275] env[65121]: WARNING openstack [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.139700] env[65121]: WARNING openstack [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.182758] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107133, 'name': ReconfigVM_Task, 'duration_secs': 0.337153} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.186598] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Reconfigured VM instance instance-0000000d to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1067.187507] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0758e6e9-eebd-40df-ad60-5824bbd560d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.196312] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.214353] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] c4096314-270a-4270-9e1d-5ace8ddbd286/c4096314-270a-4270-9e1d-5ace8ddbd286.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.215823] env[65121]: WARNING neutronclient.v2_0.client [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.216453] env[65121]: WARNING openstack [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.216790] env[65121]: WARNING openstack [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.224226] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dad7ef73-6366-47f0-8a49-898072ac4cbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.246995] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1067.246995] env[65121]: value = "task-5107134" [ 1067.246995] env[65121]: _type = "Task" [ 1067.246995] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.258536] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107134, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.321913] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "3d6fc9df-6872-4929-8813-bec1db1ebb16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.322252] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.322542] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "3d6fc9df-6872-4929-8813-bec1db1ebb16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.322896] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.323128] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.331420] env[65121]: INFO nova.compute.manager [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Terminating instance [ 1067.339037] env[65121]: DEBUG nova.compute.utils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1067.344751] env[65121]: DEBUG nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1067.345048] env[65121]: DEBUG nova.network.neutron [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1067.345390] env[65121]: WARNING neutronclient.v2_0.client [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.345701] env[65121]: WARNING neutronclient.v2_0.client [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.346949] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.347307] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.387131] env[65121]: DEBUG nova.network.neutron [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updated VIF entry in instance network info cache for port 2acd5dbc-a08c-4ebd-922b-284294369a33. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1067.387591] env[65121]: DEBUG nova.network.neutron [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1067.414464] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Releasing lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1067.415250] env[65121]: DEBUG nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Instance network_info: |[{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1067.415682] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:30:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '34a581cb-6d33-4e2e-af50-735a6749d6da', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82e421ec-a3e6-4fcb-a1f4-e88c302e1df2', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.424951] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Creating folder: Project (a98eae6244314b09865c29eff2ee56a8). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.428983] env[65121]: DEBUG nova.policy [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26beb81d28f44cc8932ea8e87a4cd2c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0630960dcbf44781be05184565d81932', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1067.434014] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-474fa45e-f9c0-4885-963f-f81c1aa055de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.446509] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107131, 'name': PowerOffVM_Task, 'duration_secs': 1.356404} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.446509] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.446509] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.446706] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b52f456-82f1-4157-92b8-a4b0871b843b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.450096] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Created folder: Project (a98eae6244314b09865c29eff2ee56a8) in parent group-v993268. [ 1067.450096] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Creating folder: Instances. Parent ref: group-v993537. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.453490] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03668cde-5a3d-4448-adb8-c523bf2f6cf8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.463911] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Created folder: Instances in parent group-v993537. [ 1067.464204] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1067.464415] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.464611] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da942e06-147a-4724-8db2-d0e7e6f18be7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.482596] env[65121]: DEBUG nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1067.492982] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.492982] env[65121]: value = "task-5107138" [ 1067.492982] env[65121]: _type = "Task" [ 1067.492982] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.503859] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107138, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.529438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.529438] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.529438] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleting the datastore file [datastore1] ec463cad-4c80-4636-bc7a-9ec298a07d96 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.529438] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d567d942-cf06-49e1-8147-edc5a896f59e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.538829] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1067.538829] env[65121]: value = "task-5107139" [ 1067.538829] env[65121]: _type = "Task" [ 1067.538829] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.558393] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.564998] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107132, 'name': PowerOffVM_Task, 'duration_secs': 0.732765} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.565960] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.566169] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.566486] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c87c131b-708b-4d45-87b4-f152dc1d4564 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.640788] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.641088] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.641646] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleting the datastore file [datastore2] abd87ca7-5d95-4b22-8d2a-1f37abd18f86 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.641730] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b4e5326-5744-4373-a9b5-e1de15107efb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.649797] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for the task: (returnval){ [ 1067.649797] env[65121]: value = "task-5107141" [ 1067.649797] env[65121]: _type = "Task" [ 1067.649797] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.663170] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107141, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.694147] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.760757] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107134, 'name': ReconfigVM_Task, 'duration_secs': 0.384686} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.760757] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Reconfigured VM instance instance-0000000d to attach disk [datastore1] c4096314-270a-4270-9e1d-5ace8ddbd286/c4096314-270a-4270-9e1d-5ace8ddbd286.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.760757] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance 'c4096314-270a-4270-9e1d-5ace8ddbd286' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.771018] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5169d43c-a0e4-4f66-9ef3-00defa37a452 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.778749] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88a00e3-65a7-4855-b935-84d457d59ef8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.814100] env[65121]: DEBUG nova.network.neutron [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Successfully created port: b7cad5ed-15a4-4f85-8d54-51dcbbd9545e {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1067.817529] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f510b7b-1dbd-490a-9730-bbf3bd4b92a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.827167] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa2fc70-56b4-4a08-9056-1d9be68dff13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.843592] env[65121]: DEBUG nova.compute.manager [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1067.843808] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.844679] env[65121]: DEBUG nova.compute.provider_tree [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1067.846503] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d01c3f0-81e6-477f-b2d1-a234faef462d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.849821] env[65121]: DEBUG nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1067.858819] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.859079] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-982e8716-3ddd-4840-9d86-3029414166d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.867653] env[65121]: DEBUG oslo_vmware.api [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1067.867653] env[65121]: value = "task-5107142" [ 1067.867653] env[65121]: _type = "Task" [ 1067.867653] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.876376] env[65121]: DEBUG oslo_vmware.api [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.894457] env[65121]: DEBUG oslo_concurrency.lockutils [req-47896dc5-05db-4280-9db0-648d9e6c156e req-a6c56966-17f3-477a-8b94-f8cf7d525c14 service nova] Releasing lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.007695] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107138, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.008807] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.053020] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.139831] env[65121]: DEBUG nova.compute.manager [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1068.140043] env[65121]: DEBUG nova.compute.manager [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing instance network info cache due to event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1068.140261] env[65121]: DEBUG oslo_concurrency.lockutils [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Acquiring lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.140397] env[65121]: DEBUG oslo_concurrency.lockutils [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Acquired lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.140547] env[65121]: DEBUG nova.network.neutron [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1068.163580] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107141, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.192415] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.268199] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27dd8c2d-ff9b-498d-8715-adf72b6debe6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.290867] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2182171d-053f-4028-876c-dfd755a8e7ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.312715] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance 'c4096314-270a-4270-9e1d-5ace8ddbd286' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1068.377045] env[65121]: ERROR nova.scheduler.client.report [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] [req-5b4104b3-810f-4a35-b170-92ac66bb3016] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d56783bf-3ede-475a-8c5a-8d8303049e47. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5b4104b3-810f-4a35-b170-92ac66bb3016"}]} [ 1068.382119] env[65121]: DEBUG oslo_vmware.api [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.392890] env[65121]: DEBUG nova.scheduler.client.report [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Refreshing inventories for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1068.409725] env[65121]: DEBUG nova.scheduler.client.report [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Updating ProviderTree inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1068.409961] env[65121]: DEBUG nova.compute.provider_tree [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1068.422493] env[65121]: DEBUG nova.scheduler.client.report [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Refreshing aggregate associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, aggregates: None {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1068.443074] env[65121]: DEBUG nova.scheduler.client.report [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Refreshing trait associations for resource provider d56783bf-3ede-475a-8c5a-8d8303049e47, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE {{(pid=65121) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1068.506546] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107138, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.558519] env[65121]: DEBUG oslo_vmware.api [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.953453} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.558831] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.559014] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.559210] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.559487] env[65121]: INFO nova.compute.manager [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Took 2.67 seconds to destroy the instance on the hypervisor. [ 1068.559885] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1068.560226] env[65121]: DEBUG nova.compute.manager [-] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1068.560348] env[65121]: DEBUG nova.network.neutron [-] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1068.560680] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.561504] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.561710] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.607378] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.644042] env[65121]: WARNING neutronclient.v2_0.client [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.644567] env[65121]: WARNING openstack [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.645129] env[65121]: WARNING openstack [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.667737] env[65121]: DEBUG oslo_vmware.api [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Task: {'id': task-5107141, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.992992} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.667992] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.668191] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.668361] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.668523] env[65121]: INFO nova.compute.manager [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Took 2.15 seconds to destroy the instance on the hypervisor. [ 1068.668754] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1068.669034] env[65121]: DEBUG nova.compute.manager [-] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1068.669240] env[65121]: DEBUG nova.network.neutron [-] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1068.669484] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.670168] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1068.670488] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1068.694183] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.810960] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e2967d-bb44-426d-bcbb-3376ef6e7e4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.820980] env[65121]: WARNING neutronclient.v2_0.client [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.824595] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37216efe-aeb6-4190-8ab1-b1ba28cfbb77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.858058] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339d8af9-f692-4ea4-9f7a-9267048df3fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.861920] env[65121]: DEBUG nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1068.875504] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbb0a8f-f0ee-4b88-ae21-1d1289c47c4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.894086] env[65121]: DEBUG oslo_vmware.api [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107142, 'name': PowerOffVM_Task, 'duration_secs': 0.898741} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.894881] env[65121]: DEBUG nova.compute.provider_tree [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1068.900017] env[65121]: DEBUG nova.virt.hardware [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1068.902356] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.902519] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.903294] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c173f1d3-5687-4b20-8661-4be4cbb43d25 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.906733] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a0fd40e-a102-4ae6-a7bd-fa4751134563 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.914685] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb2cc8a-bad0-40ca-b091-b6d3c3cd4fdc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.932498] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.983012] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.983255] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.983412] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleting the datastore file [datastore1] 3d6fc9df-6872-4929-8813-bec1db1ebb16 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.983688] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78a14345-c6ed-47b1-827c-40a9e2320cad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.990753] env[65121]: DEBUG oslo_vmware.api [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1068.990753] env[65121]: value = "task-5107144" [ 1068.990753] env[65121]: _type = "Task" [ 1068.990753] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.002639] env[65121]: DEBUG nova.network.neutron [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Port 4ea2aa90-9370-491b-97b5-2c112a535219 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1069.004029] env[65121]: DEBUG oslo_vmware.api [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107144, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.010483] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107138, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.194913] env[65121]: DEBUG oslo_vmware.api [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107129, 'name': PowerOnVM_Task, 'duration_secs': 3.558606} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.195313] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1069.195457] env[65121]: INFO nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Took 12.81 seconds to spawn the instance on the hypervisor. [ 1069.195633] env[65121]: DEBUG nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1069.196479] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c45d857-bea0-4a3a-a900-da05cf3d38f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.310818] env[65121]: DEBUG nova.compute.manager [req-9e89eab2-f327-454d-8ea6-8bc5e2871fa1 req-43130e91-62d7-4214-88b7-fd495b4d7cf0 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Received event network-vif-deleted-84a9fa7d-03b5-4acf-b07e-ae68b78b7c57 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1069.312194] env[65121]: INFO nova.compute.manager [req-9e89eab2-f327-454d-8ea6-8bc5e2871fa1 req-43130e91-62d7-4214-88b7-fd495b4d7cf0 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Neutron deleted interface 84a9fa7d-03b5-4acf-b07e-ae68b78b7c57; detaching it from the instance and deleting it from the info cache [ 1069.312194] env[65121]: DEBUG nova.network.neutron [req-9e89eab2-f327-454d-8ea6-8bc5e2871fa1 req-43130e91-62d7-4214-88b7-fd495b4d7cf0 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1069.316446] env[65121]: WARNING openstack [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.316446] env[65121]: WARNING openstack [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.362525] env[65121]: DEBUG nova.network.neutron [-] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1069.368820] env[65121]: DEBUG nova.network.neutron [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Successfully updated port: b7cad5ed-15a4-4f85-8d54-51dcbbd9545e {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1069.399007] env[65121]: WARNING neutronclient.v2_0.client [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1069.399843] env[65121]: WARNING openstack [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.400209] env[65121]: WARNING openstack [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.445239] env[65121]: DEBUG nova.scheduler.client.report [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Updated inventory for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with generation 134 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1069.445520] env[65121]: DEBUG nova.compute.provider_tree [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Updating resource provider d56783bf-3ede-475a-8c5a-8d8303049e47 generation from 134 to 135 during operation: update_inventory {{(pid=65121) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1069.445698] env[65121]: DEBUG nova.compute.provider_tree [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Updating inventory in ProviderTree for provider d56783bf-3ede-475a-8c5a-8d8303049e47 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1069.511363] env[65121]: DEBUG oslo_vmware.api [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107144, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.40432} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.514251] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.514439] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.517017] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.517017] env[65121]: INFO nova.compute.manager [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Took 1.67 seconds to destroy the instance on the hypervisor. [ 1069.517017] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1069.517017] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107138, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.517017] env[65121]: DEBUG nova.compute.manager [-] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1069.517017] env[65121]: DEBUG nova.network.neutron [-] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1069.517017] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1069.517017] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.517017] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.526028] env[65121]: DEBUG nova.network.neutron [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updated VIF entry in instance network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1069.526383] env[65121]: DEBUG nova.network.neutron [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1069.568319] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1069.716345] env[65121]: INFO nova.compute.manager [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Took 36.35 seconds to build instance. [ 1069.726142] env[65121]: DEBUG nova.network.neutron [-] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1069.825660] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37205583-1f6f-4126-8c13-00860c071102 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.836768] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2764eed8-f2b9-40cc-8fd8-af22644ba373 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.878603] env[65121]: INFO nova.compute.manager [-] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Took 1.32 seconds to deallocate network for instance. [ 1069.879112] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-97fa5791-28cb-4e77-bad7-f2cbce8bd06d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.879278] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-97fa5791-28cb-4e77-bad7-f2cbce8bd06d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1069.879450] env[65121]: DEBUG nova.network.neutron [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1069.880698] env[65121]: DEBUG nova.compute.manager [req-9e89eab2-f327-454d-8ea6-8bc5e2871fa1 req-43130e91-62d7-4214-88b7-fd495b4d7cf0 service nova] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Detach interface failed, port_id=84a9fa7d-03b5-4acf-b07e-ae68b78b7c57, reason: Instance ec463cad-4c80-4636-bc7a-9ec298a07d96 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1069.952590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.117s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.955320] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.965s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.955595] env[65121]: DEBUG nova.objects.instance [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'resources' on Instance uuid 23d6f632-8918-46af-b239-08a9615dfbec {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.984385] env[65121]: INFO nova.scheduler.client.report [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Deleted allocations for instance 11caa171-0314-4e7d-8502-80932b469da8 [ 1070.012035] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107138, 'name': CreateVM_Task, 'duration_secs': 2.308063} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.012035] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1070.012035] env[65121]: WARNING neutronclient.v2_0.client [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.012035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.012035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1070.012035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1070.012035] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d83f2344-f334-4377-9224-d4f20da6d2f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.025772] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.026017] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.026198] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.038164] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1070.038164] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523ca60d-e28b-7355-1b1f-2c7e99ab5fe6" [ 1070.038164] env[65121]: _type = "Task" [ 1070.038164] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.041589] env[65121]: DEBUG oslo_concurrency.lockutils [req-3fb064ce-8f9a-4c65-9ee8-4f6405221c5e req-cb9ad6ed-dbe5-4317-a8d7-e10ddcebb74b service nova] Releasing lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.056235] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523ca60d-e28b-7355-1b1f-2c7e99ab5fe6, 'name': SearchDatastore_Task, 'duration_secs': 0.013544} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.057815] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.057815] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.057815] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.057815] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1070.057815] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.058414] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc74fe9c-b907-41e2-a87a-beac9bc4dc90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.075442] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.075570] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1070.076338] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11409443-0b5b-4ccf-9c54-d85295b5b7bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.082827] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1070.082827] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528e1679-b482-ba1a-e551-6af15ff6bbee" [ 1070.082827] env[65121]: _type = "Task" [ 1070.082827] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.093945] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528e1679-b482-ba1a-e551-6af15ff6bbee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.172697] env[65121]: DEBUG nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Received event network-vif-plugged-b7cad5ed-15a4-4f85-8d54-51dcbbd9545e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1070.172927] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Acquiring lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.173143] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1070.173303] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.173460] env[65121]: DEBUG nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] No waiting events found dispatching network-vif-plugged-b7cad5ed-15a4-4f85-8d54-51dcbbd9545e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1070.173615] env[65121]: WARNING nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Received unexpected event network-vif-plugged-b7cad5ed-15a4-4f85-8d54-51dcbbd9545e for instance with vm_state building and task_state spawning. [ 1070.173790] env[65121]: DEBUG nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Received event network-changed-b7cad5ed-15a4-4f85-8d54-51dcbbd9545e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1070.173952] env[65121]: DEBUG nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Refreshing instance network info cache due to event network-changed-b7cad5ed-15a4-4f85-8d54-51dcbbd9545e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1070.174124] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Acquiring lock "refresh_cache-97fa5791-28cb-4e77-bad7-f2cbce8bd06d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.218519] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3118a02f-3352-4a62-81fc-ba338c489f04 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "fc3109cb-72bb-4695-82c3-a323e619919d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.864s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.228295] env[65121]: INFO nova.compute.manager [-] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Took 1.56 seconds to deallocate network for instance. [ 1070.299713] env[65121]: DEBUG nova.network.neutron [-] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1070.385278] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.385278] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.393840] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.431508] env[65121]: DEBUG nova.network.neutron [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1070.484516] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.484924] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.499531] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b49e014f-9f5f-438a-9717-ca46bfe1bbde tempest-ServerDiagnosticsTest-28763545 tempest-ServerDiagnosticsTest-28763545-project-member] Lock "11caa171-0314-4e7d-8502-80932b469da8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.885s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.578177] env[65121]: WARNING neutronclient.v2_0.client [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.578912] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.580555] env[65121]: WARNING openstack [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.602017] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528e1679-b482-ba1a-e551-6af15ff6bbee, 'name': SearchDatastore_Task, 'duration_secs': 0.016736} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.602849] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9acd4ef6-a6eb-40aa-ac4f-128e0ca1fc4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.609861] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1070.609861] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e1f312-73ba-11a8-74ec-97429bc0698f" [ 1070.609861] env[65121]: _type = "Task" [ 1070.609861] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.618764] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e1f312-73ba-11a8-74ec-97429bc0698f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.698431] env[65121]: DEBUG nova.network.neutron [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Updating instance_info_cache with network_info: [{"id": "b7cad5ed-15a4-4f85-8d54-51dcbbd9545e", "address": "fa:16:3e:41:73:2a", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7cad5ed-15", "ovs_interfaceid": "b7cad5ed-15a4-4f85-8d54-51dcbbd9545e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1070.738557] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.803267] env[65121]: INFO nova.compute.manager [-] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Took 1.29 seconds to deallocate network for instance. [ 1070.859882] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b6545b-b659-4414-97e5-c479b387e3eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.869449] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2780910-8688-4d89-a861-77a517edbde2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.904516] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b4f457-8779-466e-926c-a2c872d376b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.913955] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8f4df3-2db8-41ff-b499-e99845bf4b77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.931431] env[65121]: DEBUG nova.compute.provider_tree [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.046297] env[65121]: WARNING neutronclient.v2_0.client [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.079144] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.079144] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.079144] env[65121]: DEBUG nova.network.neutron [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1071.123165] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e1f312-73ba-11a8-74ec-97429bc0698f, 'name': SearchDatastore_Task, 'duration_secs': 0.038404} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.123444] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1071.123686] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/03a13029-55b1-4757-80a6-9293471434dc.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1071.123954] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-594676da-2d78-4e78-b749-5d11042cb4db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.131307] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1071.131307] env[65121]: value = "task-5107145" [ 1071.131307] env[65121]: _type = "Task" [ 1071.131307] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.140396] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107145, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.202878] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-97fa5791-28cb-4e77-bad7-f2cbce8bd06d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1071.203401] env[65121]: DEBUG nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Instance network_info: |[{"id": "b7cad5ed-15a4-4f85-8d54-51dcbbd9545e", "address": "fa:16:3e:41:73:2a", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7cad5ed-15", "ovs_interfaceid": "b7cad5ed-15a4-4f85-8d54-51dcbbd9545e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1071.203862] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Acquired lock "refresh_cache-97fa5791-28cb-4e77-bad7-f2cbce8bd06d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.204117] env[65121]: DEBUG nova.network.neutron [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Refreshing network info cache for port b7cad5ed-15a4-4f85-8d54-51dcbbd9545e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1071.205390] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:73:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7cad5ed-15a4-4f85-8d54-51dcbbd9545e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.213721] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1071.215162] env[65121]: WARNING neutronclient.v2_0.client [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.215894] env[65121]: WARNING openstack [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.216314] env[65121]: WARNING openstack [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.224011] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1071.224663] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0eef793b-37ae-46eb-a97d-46adf980684c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.248220] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.248220] env[65121]: value = "task-5107146" [ 1071.248220] env[65121]: _type = "Task" [ 1071.248220] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.260038] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107146, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.313986] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.367398] env[65121]: WARNING openstack [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.367926] env[65121]: WARNING openstack [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.383820] env[65121]: DEBUG nova.compute.manager [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Received event network-vif-deleted-6d471ad5-5dee-4d1a-8f3d-ab1235cb1d35 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1071.383961] env[65121]: DEBUG nova.compute.manager [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Received event network-changed-65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1071.384313] env[65121]: DEBUG nova.compute.manager [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Refreshing instance network info cache due to event network-changed-65c052c9-c53e-40b1-a26b-6b8e80093cee. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1071.384377] env[65121]: DEBUG oslo_concurrency.lockutils [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Acquiring lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.384496] env[65121]: DEBUG oslo_concurrency.lockutils [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Acquired lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.384652] env[65121]: DEBUG nova.network.neutron [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Refreshing network info cache for port 65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1071.435561] env[65121]: DEBUG nova.scheduler.client.report [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1071.496781] env[65121]: WARNING neutronclient.v2_0.client [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.497597] env[65121]: WARNING openstack [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.498077] env[65121]: WARNING openstack [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.581944] env[65121]: WARNING neutronclient.v2_0.client [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.583104] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.583669] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.639346] env[65121]: DEBUG nova.network.neutron [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Updated VIF entry in instance network info cache for port b7cad5ed-15a4-4f85-8d54-51dcbbd9545e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1071.640555] env[65121]: DEBUG nova.network.neutron [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Updating instance_info_cache with network_info: [{"id": "b7cad5ed-15a4-4f85-8d54-51dcbbd9545e", "address": "fa:16:3e:41:73:2a", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7cad5ed-15", "ovs_interfaceid": "b7cad5ed-15a4-4f85-8d54-51dcbbd9545e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1071.655939] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107145, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.752105] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.752105] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.769586] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107146, 'name': CreateVM_Task, 'duration_secs': 0.404796} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.770213] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.771318] env[65121]: WARNING neutronclient.v2_0.client [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.772017] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.772325] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.772931] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1071.773339] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-faa5549a-c2df-47c4-b67a-a3704739bd44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.779612] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1071.779612] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52812f9f-273f-957f-fa54-aca0be88cd6e" [ 1071.779612] env[65121]: _type = "Task" [ 1071.779612] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.789843] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52812f9f-273f-957f-fa54-aca0be88cd6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.841876] env[65121]: WARNING neutronclient.v2_0.client [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.842559] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.842897] env[65121]: WARNING openstack [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.889918] env[65121]: WARNING neutronclient.v2_0.client [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.891434] env[65121]: WARNING openstack [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.891822] env[65121]: WARNING openstack [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.944399] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.989s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1071.948630] env[65121]: DEBUG nova.network.neutron [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1071.952196] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.141s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.986981] env[65121]: INFO nova.scheduler.client.report [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted allocations for instance 23d6f632-8918-46af-b239-08a9615dfbec [ 1072.064152] env[65121]: WARNING openstack [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.064152] env[65121]: WARNING openstack [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.138702] env[65121]: WARNING neutronclient.v2_0.client [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.139448] env[65121]: WARNING openstack [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.139876] env[65121]: WARNING openstack [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.148689] env[65121]: DEBUG oslo_concurrency.lockutils [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] Releasing lock "refresh_cache-97fa5791-28cb-4e77-bad7-f2cbce8bd06d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1072.148938] env[65121]: DEBUG nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Received event network-vif-deleted-c73a5f2e-adbf-411c-bec3-31dd92e45737 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1072.149124] env[65121]: INFO nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Neutron deleted interface c73a5f2e-adbf-411c-bec3-31dd92e45737; detaching it from the instance and deleting it from the info cache [ 1072.149444] env[65121]: DEBUG nova.network.neutron [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1072.157347] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107145, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.686543} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.157600] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/03a13029-55b1-4757-80a6-9293471434dc.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1072.157812] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.158121] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70beda9c-5c7a-44e8-803d-eb1d8b201b02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.167077] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1072.167077] env[65121]: value = "task-5107147" [ 1072.167077] env[65121]: _type = "Task" [ 1072.167077] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.177701] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107147, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.229809] env[65121]: DEBUG nova.network.neutron [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Updated VIF entry in instance network info cache for port 65c052c9-c53e-40b1-a26b-6b8e80093cee. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1072.230237] env[65121]: DEBUG nova.network.neutron [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Updating instance_info_cache with network_info: [{"id": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "address": "fa:16:3e:34:30:0f", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65c052c9-c5", "ovs_interfaceid": "65c052c9-c53e-40b1-a26b-6b8e80093cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1072.292797] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52812f9f-273f-957f-fa54-aca0be88cd6e, 'name': SearchDatastore_Task, 'duration_secs': 0.036379} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.293115] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1072.293360] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.293637] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.293778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1072.293971] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.294282] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bec220f7-56b8-4902-926f-189aa229c7b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.305305] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.305468] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1072.306322] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36dfe673-37d8-4b8a-895d-cc5571347139 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.314135] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1072.314135] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5246d8d0-33c2-fb1f-5bb2-3039b3e73f80" [ 1072.314135] env[65121]: _type = "Task" [ 1072.314135] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.325470] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246d8d0-33c2-fb1f-5bb2-3039b3e73f80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.452772] env[65121]: DEBUG oslo_concurrency.lockutils [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1072.497291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1ce26c0c-963b-4c9f-9d9d-afc83f2a358d tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "23d6f632-8918-46af-b239-08a9615dfbec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.146s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.653479] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-610ff388-21a6-453d-bdce-e202ba6918d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.664316] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f5aeac-96b8-4f82-b6f1-c59110581c22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.686794] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107147, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142913} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.690707] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.692953] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad02909-b1c8-445f-be70-30766a84cf5e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.711096] env[65121]: DEBUG nova.compute.manager [req-5f0ed8ac-21ec-44fc-b9d8-cb07888a28a4 req-d260e746-4e95-4c9f-bbb1-ab2ddb227dbc service nova] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Detach interface failed, port_id=c73a5f2e-adbf-411c-bec3-31dd92e45737, reason: Instance 3d6fc9df-6872-4929-8813-bec1db1ebb16 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1072.734380] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/03a13029-55b1-4757-80a6-9293471434dc.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.737587] env[65121]: DEBUG oslo_concurrency.lockutils [req-50a80cb7-0c96-48b3-921e-be56c3a03a5d req-4101895c-e8cb-4bf5-89ed-acdedd0893b1 service nova] Releasing lock "refresh_cache-fc3109cb-72bb-4695-82c3-a323e619919d" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1072.737898] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46afb388-be30-4a48-b2ea-facd90053944 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.759961] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1072.759961] env[65121]: value = "task-5107148" [ 1072.759961] env[65121]: _type = "Task" [ 1072.759961] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.773437] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107148, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.804596] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3a0213-a1cb-4b6a-9adb-fc9f28ec7661 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.814842] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f3f33a-8021-4480-a385-32d233087015 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.827771] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246d8d0-33c2-fb1f-5bb2-3039b3e73f80, 'name': SearchDatastore_Task, 'duration_secs': 0.026048} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.857929] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78818c38-fbc5-4156-9eb9-60b5538deaf6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.860991] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df3d107-7e0f-461e-ad51-8aefa77e1bff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.869712] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1072.869712] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c0627b-8f34-e34b-255b-2a77594ce9ed" [ 1072.869712] env[65121]: _type = "Task" [ 1072.869712] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.870920] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67758da-4a6e-4c7e-8a67-31388ab6db12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.892996] env[65121]: DEBUG nova.compute.provider_tree [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.894804] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c0627b-8f34-e34b-255b-2a77594ce9ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.979978] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d3bcf3-ed7c-4fb3-aec2-c78d39132b92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.000403] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a3f257-3197-4fae-9e8f-f0eeb8492f05 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.009616] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance 'c4096314-270a-4270-9e1d-5ace8ddbd286' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1073.272571] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.388740] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c0627b-8f34-e34b-255b-2a77594ce9ed, 'name': SearchDatastore_Task, 'duration_secs': 0.023346} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.389109] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1073.389413] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 97fa5791-28cb-4e77-bad7-f2cbce8bd06d/97fa5791-28cb-4e77-bad7-f2cbce8bd06d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1073.389632] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-011a8746-7077-4c91-859b-119e659e92ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.396585] env[65121]: DEBUG nova.scheduler.client.report [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1073.403847] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1073.403847] env[65121]: value = "task-5107149" [ 1073.403847] env[65121]: _type = "Task" [ 1073.403847] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.415237] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107149, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.519906] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.520197] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-074e527a-f3d8-47f4-b7dc-65d6ddd849c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.528514] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1073.528514] env[65121]: value = "task-5107150" [ 1073.528514] env[65121]: _type = "Task" [ 1073.528514] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.537607] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107150, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.771951] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107148, 'name': ReconfigVM_Task, 'duration_secs': 0.889227} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.772320] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/03a13029-55b1-4757-80a6-9293471434dc.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.773019] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-daf5cf89-a18d-4c54-9b07-0ceae5343b2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.781692] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1073.781692] env[65121]: value = "task-5107151" [ 1073.781692] env[65121]: _type = "Task" [ 1073.781692] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.795405] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107151, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.923328] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107149, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.043403] env[65121]: DEBUG oslo_vmware.api [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107150, 'name': PowerOnVM_Task, 'duration_secs': 0.462358} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.044347] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.044347] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-441877cb-c608-4375-98d2-d84ed34dbbdc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance 'c4096314-270a-4270-9e1d-5ace8ddbd286' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1074.293700] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107151, 'name': Rename_Task, 'duration_secs': 0.378967} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.294042] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.294385] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6795a425-f433-441e-94f0-ab8d65ce8a67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.302887] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1074.302887] env[65121]: value = "task-5107152" [ 1074.302887] env[65121]: _type = "Task" [ 1074.302887] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.313064] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.417433] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.467s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.424757] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.599s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.430635] env[65121]: INFO nova.compute.claims [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.433881] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107149, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.650719} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.434618] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 97fa5791-28cb-4e77-bad7-f2cbce8bd06d/97fa5791-28cb-4e77-bad7-f2cbce8bd06d.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1074.434875] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1074.435413] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b96220d4-3d99-482c-91ae-519a88adadc3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.444985] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1074.444985] env[65121]: value = "task-5107153" [ 1074.444985] env[65121]: _type = "Task" [ 1074.444985] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.463320] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107153, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.814683] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107152, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.957642] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107153, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082416} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.957642] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1074.958886] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddce4e2-6fcf-48ad-aec5-088c9cc31e05 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.988635] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 97fa5791-28cb-4e77-bad7-f2cbce8bd06d/97fa5791-28cb-4e77-bad7-f2cbce8bd06d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.989031] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d337b765-041a-49be-9591-3e34d6b08cfa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.013932] env[65121]: INFO nova.scheduler.client.report [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted allocation for migration aeab73c9-8ca4-4e9d-82d4-d10914fed2b0 [ 1075.016311] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1075.016311] env[65121]: value = "task-5107154" [ 1075.016311] env[65121]: _type = "Task" [ 1075.016311] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.029224] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107154, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.315305] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107152, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.525340] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f6d1e0ee-8a0f-4636-87ab-e6970de858b6 tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.212s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.535492] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107154, 'name': ReconfigVM_Task, 'duration_secs': 0.3265} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.535819] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 97fa5791-28cb-4e77-bad7-f2cbce8bd06d/97fa5791-28cb-4e77-bad7-f2cbce8bd06d.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1075.536865] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6388a62-be1e-4cd2-8e3d-435ec334916f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.545727] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1075.545727] env[65121]: value = "task-5107155" [ 1075.545727] env[65121]: _type = "Task" [ 1075.545727] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.557605] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107155, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.610155] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.610934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.610934] env[65121]: INFO nova.compute.manager [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Shelving [ 1075.628049] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.628301] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.814593] env[65121]: DEBUG oslo_vmware.api [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107152, 'name': PowerOnVM_Task, 'duration_secs': 1.226868} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.814839] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1075.815527] env[65121]: INFO nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Took 10.86 seconds to spawn the instance on the hypervisor. [ 1075.815527] env[65121]: DEBUG nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1075.816029] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d49e200-1042-4dc5-9fd6-984e172fe900 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.820278] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8915ca9a-3174-4488-baf6-4f13bbea9ac7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.829205] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab000c11-69a8-4a43-a985-dc8035ef72cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.872370] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d72851-6e86-4ddc-a3af-0a9e285f3c92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.883107] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfb12ec-4967-4a21-ae10-6de854b65a09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.898817] env[65121]: DEBUG nova.compute.provider_tree [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.056322] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107155, 'name': Rename_Task, 'duration_secs': 0.148684} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.056613] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1076.056868] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02ee3e07-0bec-42c9-94c1-fe9b4bbdb9fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.065463] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1076.065463] env[65121]: value = "task-5107156" [ 1076.065463] env[65121]: _type = "Task" [ 1076.065463] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.076751] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.130443] env[65121]: DEBUG nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1076.211146] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.211536] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.211826] env[65121]: DEBUG nova.compute.manager [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Going to confirm migration 5 {{(pid=65121) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1076.346196] env[65121]: INFO nova.compute.manager [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Took 17.84 seconds to build instance. [ 1076.401906] env[65121]: DEBUG nova.scheduler.client.report [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1076.576618] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107156, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.622949] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1076.622949] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7e86657-cb43-4482-b7a1-a8da88a374c5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.632528] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1076.632528] env[65121]: value = "task-5107157" [ 1076.632528] env[65121]: _type = "Task" [ 1076.632528] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.647996] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.660027] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.690189] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "564159fe-6ecd-4276-8d65-cbe25859493a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.690471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.690801] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.690994] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.691194] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.693632] env[65121]: INFO nova.compute.manager [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Terminating instance [ 1076.719946] env[65121]: WARNING neutronclient.v2_0.client [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.760780] env[65121]: WARNING neutronclient.v2_0.client [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.761472] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.761711] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.761905] env[65121]: DEBUG nova.network.neutron [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1076.762103] env[65121]: DEBUG nova.objects.instance [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'info_cache' on Instance uuid c4096314-270a-4270-9e1d-5ace8ddbd286 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.848443] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3b188e5b-159d-450e-bf13-85c7ddbfeb50 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "03a13029-55b1-4757-80a6-9293471434dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.347s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.907863] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.908514] env[65121]: DEBUG nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1076.911405] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.902s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.912480] env[65121]: INFO nova.compute.claims [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.078456] env[65121]: DEBUG oslo_vmware.api [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107156, 'name': PowerOnVM_Task, 'duration_secs': 0.517309} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.078763] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1077.079029] env[65121]: INFO nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Took 8.22 seconds to spawn the instance on the hypervisor. [ 1077.079276] env[65121]: DEBUG nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1077.080326] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c78f0be-d87e-4f5e-a442-26e359d00a3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.145869] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107157, 'name': PowerOffVM_Task, 'duration_secs': 0.504392} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.146517] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1077.147472] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f85f3f-5cc4-4642-8174-6569ef827ee2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.168956] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e47e616-56b3-477b-ac1d-4c6164c4f220 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.198487] env[65121]: DEBUG nova.compute.manager [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1077.198750] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1077.199745] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4aa2027-602e-467b-995b-0b7af949cbc2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.208946] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1077.209234] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5ba8244-df9c-4f38-8e9b-f6995c58003a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.216560] env[65121]: DEBUG oslo_vmware.api [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1077.216560] env[65121]: value = "task-5107158" [ 1077.216560] env[65121]: _type = "Task" [ 1077.216560] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.227565] env[65121]: DEBUG oslo_vmware.api [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.244600] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1077.244897] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1077.315902] env[65121]: INFO nova.compute.manager [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Rescuing [ 1077.316201] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.316360] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquired lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.316528] env[65121]: DEBUG nova.network.neutron [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1077.416940] env[65121]: DEBUG nova.compute.utils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1077.420833] env[65121]: DEBUG nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1077.421088] env[65121]: DEBUG nova.network.neutron [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1077.421421] env[65121]: WARNING neutronclient.v2_0.client [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.421742] env[65121]: WARNING neutronclient.v2_0.client [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.422395] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.422759] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.490170] env[65121]: DEBUG nova.policy [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '54a5ae8431a9432ebe019f75b26d1a17', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '665a22b60ef542198ef02de9d6439807', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1077.599618] env[65121]: INFO nova.compute.manager [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Took 14.67 seconds to build instance. [ 1077.682189] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1077.682823] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6722b22b-8283-4428-a310-753963cba81c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.692027] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1077.692027] env[65121]: value = "task-5107159" [ 1077.692027] env[65121]: _type = "Task" [ 1077.692027] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.702264] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107159, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.730505] env[65121]: DEBUG oslo_vmware.api [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107158, 'name': PowerOffVM_Task, 'duration_secs': 0.360491} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.730822] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1077.730955] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1077.731229] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ada823d8-8959-40aa-8ebb-cab8e72be4d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.748695] env[65121]: DEBUG nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1077.768721] env[65121]: WARNING neutronclient.v2_0.client [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.769863] env[65121]: WARNING openstack [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.769990] env[65121]: WARNING openstack [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.801199] env[65121]: DEBUG nova.network.neutron [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Successfully created port: 7fe57d90-496e-44ce-bab8-1f328075e76b {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1077.808822] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1077.809144] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1077.809380] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleting the datastore file [datastore1] 564159fe-6ecd-4276-8d65-cbe25859493a {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.809666] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfc3706a-f1b6-46b4-a803-1b84b0c624a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.818275] env[65121]: DEBUG oslo_vmware.api [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1077.818275] env[65121]: value = "task-5107161" [ 1077.818275] env[65121]: _type = "Task" [ 1077.818275] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.822146] env[65121]: WARNING neutronclient.v2_0.client [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.822932] env[65121]: WARNING openstack [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.823454] env[65121]: WARNING openstack [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.843438] env[65121]: DEBUG oslo_vmware.api [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.921583] env[65121]: DEBUG nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1078.028666] env[65121]: WARNING openstack [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.029629] env[65121]: WARNING openstack [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.063927] env[65121]: WARNING openstack [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.064359] env[65121]: WARNING openstack [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.105993] env[65121]: DEBUG oslo_concurrency.lockutils [None req-143e77ef-41f9-41d3-bde5-0ac8d1ed1bcc tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.913s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.157918] env[65121]: DEBUG oslo_concurrency.lockutils [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1078.158205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1078.158386] env[65121]: DEBUG nova.compute.manager [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1078.162235] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a783a32-48fa-4208-984e-741bb0562381 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.166191] env[65121]: WARNING neutronclient.v2_0.client [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.166908] env[65121]: WARNING openstack [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.167418] env[65121]: WARNING openstack [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.182365] env[65121]: DEBUG nova.compute.manager [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1078.183317] env[65121]: DEBUG nova.objects.instance [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'flavor' on Instance uuid 97fa5791-28cb-4e77-bad7-f2cbce8bd06d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.207158] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107159, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.261256] env[65121]: WARNING neutronclient.v2_0.client [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.262165] env[65121]: WARNING openstack [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.263542] env[65121]: WARNING openstack [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.278291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1078.332178] env[65121]: DEBUG oslo_vmware.api [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173969} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.332451] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1078.332652] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1078.332916] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1078.333592] env[65121]: INFO nova.compute.manager [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1078.333592] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1078.333592] env[65121]: DEBUG nova.compute.manager [-] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1078.333752] env[65121]: DEBUG nova.network.neutron [-] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1078.333914] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.334454] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.334756] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.343604] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdd1350-af55-4acb-a4eb-47bfc9db3d7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.352798] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa5880e-742e-402a-9a14-7822b51979ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.391886] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8a44bd-ddd7-40b0-943d-d6270b89284b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.400954] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1ba3e7-8a17-4c08-8a1a-93c905ee1f3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.416272] env[65121]: DEBUG nova.compute.provider_tree [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.419882] env[65121]: DEBUG nova.network.neutron [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [{"id": "4ea2aa90-9370-491b-97b5-2c112a535219", "address": "fa:16:3e:a1:95:0d", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ea2aa90-93", "ovs_interfaceid": "4ea2aa90-9370-491b-97b5-2c112a535219", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.494103] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.591571] env[65121]: DEBUG nova.network.neutron [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.706668] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107159, 'name': CreateSnapshot_Task, 'duration_secs': 0.526346} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.706971] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1078.707932] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ff76a0-33f2-4df0-a8ad-30cfd0e19fbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.807145] env[65121]: DEBUG nova.compute.manager [req-247c2563-e1bb-431d-b388-c0eda350e652 req-8c816c76-999d-4ff5-83a6-762e24c4190a service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Received event network-vif-deleted-db0cba48-a3d6-4aff-b44f-c1e6672a1daf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1078.807454] env[65121]: INFO nova.compute.manager [req-247c2563-e1bb-431d-b388-c0eda350e652 req-8c816c76-999d-4ff5-83a6-762e24c4190a service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Neutron deleted interface db0cba48-a3d6-4aff-b44f-c1e6672a1daf; detaching it from the instance and deleting it from the info cache [ 1078.807644] env[65121]: DEBUG nova.network.neutron [req-247c2563-e1bb-431d-b388-c0eda350e652 req-8c816c76-999d-4ff5-83a6-762e24c4190a service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.922930] env[65121]: DEBUG nova.scheduler.client.report [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.926494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-c4096314-270a-4270-9e1d-5ace8ddbd286" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.927278] env[65121]: DEBUG nova.objects.instance [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'migration_context' on Instance uuid c4096314-270a-4270-9e1d-5ace8ddbd286 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.932929] env[65121]: DEBUG nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1078.960819] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1078.961077] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1078.961228] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1078.961407] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1078.961545] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1078.961699] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1078.961903] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.962079] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1078.962249] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1078.962407] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1078.962571] env[65121]: DEBUG nova.virt.hardware [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1078.963463] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b2dd3b-4da2-489f-ba30-22f8058a136d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.972851] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2df264-15b3-4115-893c-12279654f7a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.094982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Releasing lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.194244] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.194582] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-603bdb2f-9592-484b-85ab-9447e92342bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.203443] env[65121]: DEBUG oslo_vmware.api [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1079.203443] env[65121]: value = "task-5107162" [ 1079.203443] env[65121]: _type = "Task" [ 1079.203443] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.212489] env[65121]: DEBUG oslo_vmware.api [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.229381] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1079.230583] env[65121]: DEBUG nova.compute.manager [req-59821fda-a897-46b4-92ad-1513ff90fa42 req-74412ed0-c492-41fb-b494-6107df3fd77f service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Received event network-vif-plugged-7fe57d90-496e-44ce-bab8-1f328075e76b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1079.230780] env[65121]: DEBUG oslo_concurrency.lockutils [req-59821fda-a897-46b4-92ad-1513ff90fa42 req-74412ed0-c492-41fb-b494-6107df3fd77f service nova] Acquiring lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.230987] env[65121]: DEBUG oslo_concurrency.lockutils [req-59821fda-a897-46b4-92ad-1513ff90fa42 req-74412ed0-c492-41fb-b494-6107df3fd77f service nova] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.231161] env[65121]: DEBUG oslo_concurrency.lockutils [req-59821fda-a897-46b4-92ad-1513ff90fa42 req-74412ed0-c492-41fb-b494-6107df3fd77f service nova] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.231440] env[65121]: DEBUG nova.compute.manager [req-59821fda-a897-46b4-92ad-1513ff90fa42 req-74412ed0-c492-41fb-b494-6107df3fd77f service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] No waiting events found dispatching network-vif-plugged-7fe57d90-496e-44ce-bab8-1f328075e76b {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1079.231708] env[65121]: WARNING nova.compute.manager [req-59821fda-a897-46b4-92ad-1513ff90fa42 req-74412ed0-c492-41fb-b494-6107df3fd77f service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Received unexpected event network-vif-plugged-7fe57d90-496e-44ce-bab8-1f328075e76b for instance with vm_state building and task_state spawning. [ 1079.232420] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-24c9aa16-6fc7-440b-9fd7-e8302fb4796d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.242139] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1079.242139] env[65121]: value = "task-5107163" [ 1079.242139] env[65121]: _type = "Task" [ 1079.242139] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.252283] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107163, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.283121] env[65121]: DEBUG nova.network.neutron [-] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1079.310807] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86fcc4c3-ec93-40cd-a8e6-63467b9d835a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.322340] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ba84ef-2cd3-4b62-8fa5-c2f69d9f0637 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.346486] env[65121]: DEBUG nova.network.neutron [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Successfully updated port: 7fe57d90-496e-44ce-bab8-1f328075e76b {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1079.363444] env[65121]: DEBUG nova.compute.manager [req-247c2563-e1bb-431d-b388-c0eda350e652 req-8c816c76-999d-4ff5-83a6-762e24c4190a service nova] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Detach interface failed, port_id=db0cba48-a3d6-4aff-b44f-c1e6672a1daf, reason: Instance 564159fe-6ecd-4276-8d65-cbe25859493a could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1079.429893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.430521] env[65121]: DEBUG nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1079.434594] env[65121]: DEBUG nova.objects.base [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1079.435475] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.042s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.435566] env[65121]: DEBUG nova.objects.instance [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lazy-loading 'resources' on Instance uuid ec463cad-4c80-4636-bc7a-9ec298a07d96 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.439076] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83daa3b2-d494-44d9-9964-3255e5293b97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.461405] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0640e1b-9085-408a-b663-8099d3e0ae84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.467853] env[65121]: DEBUG oslo_vmware.api [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1079.467853] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5256d765-7a77-803b-2ec2-d87de2a92173" [ 1079.467853] env[65121]: _type = "Task" [ 1079.467853] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.479599] env[65121]: DEBUG oslo_vmware.api [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5256d765-7a77-803b-2ec2-d87de2a92173, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.714872] env[65121]: DEBUG oslo_vmware.api [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107162, 'name': PowerOffVM_Task, 'duration_secs': 0.422253} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.715187] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.715430] env[65121]: DEBUG nova.compute.manager [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1079.716274] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde6765e-a7c3-4146-a0a2-d5a82331b603 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.753525] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107163, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.786681] env[65121]: INFO nova.compute.manager [-] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Took 1.45 seconds to deallocate network for instance. [ 1079.864901] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "refresh_cache-347bbb23-da89-4e6f-837c-e88bf7c57a3a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.865218] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquired lock "refresh_cache-347bbb23-da89-4e6f-837c-e88bf7c57a3a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.865429] env[65121]: DEBUG nova.network.neutron [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1079.937279] env[65121]: DEBUG nova.compute.utils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1079.938901] env[65121]: DEBUG nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Not allocating networking since 'none' was specified. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1079.986652] env[65121]: DEBUG oslo_vmware.api [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5256d765-7a77-803b-2ec2-d87de2a92173, 'name': SearchDatastore_Task, 'duration_secs': 0.012104} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.986935] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1080.231145] env[65121]: DEBUG oslo_concurrency.lockutils [None req-02f88171-39e6-4724-a6f7-abea74c5e212 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.073s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1080.254759] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107163, 'name': CloneVM_Task} progress is 95%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.288264] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcdf30e-b191-4eec-8659-b18c11b7958c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.295455] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1080.301140] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c95a42-10c9-4206-bbd2-e8381c544d9b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.344387] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a51e52-16f1-4a8b-b09f-e5de81280f22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.354956] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e4bdc3-83b2-4c60-9d4c-e2887288b164 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.372568] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.373040] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.381346] env[65121]: DEBUG nova.compute.provider_tree [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.413439] env[65121]: DEBUG nova.network.neutron [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1080.433379] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.433773] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.442818] env[65121]: DEBUG nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1080.501922] env[65121]: WARNING neutronclient.v2_0.client [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.502707] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.503149] env[65121]: WARNING openstack [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.628572] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.631098] env[65121]: DEBUG nova.network.neutron [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Updating instance_info_cache with network_info: [{"id": "7fe57d90-496e-44ce-bab8-1f328075e76b", "address": "fa:16:3e:c5:e6:c5", "network": {"id": "7293d35b-5f53-4ff1-bb9f-a4724377a87b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-972919019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "665a22b60ef542198ef02de9d6439807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69076131-87ac-46dd-9d5d-8d1b4ea7dec6", "external-id": "nsx-vlan-transportzone-327", "segmentation_id": 327, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fe57d90-49", "ovs_interfaceid": "7fe57d90-496e-44ce-bab8-1f328075e76b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1080.632258] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22014a3b-8c78-490a-82f0-f1890a6c4736 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.640137] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1080.640137] env[65121]: value = "task-5107164" [ 1080.640137] env[65121]: _type = "Task" [ 1080.640137] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.650180] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.753948] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107163, 'name': CloneVM_Task, 'duration_secs': 1.135333} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.754232] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Created linked-clone VM from snapshot [ 1080.755204] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6ae007-6eb4-43e3-a81f-c78cb831ac88 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.762958] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Uploading image 872a03ef-a1b4-487f-afd6-17a5c0d28e41 {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1080.788615] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1080.788615] env[65121]: value = "vm-993542" [ 1080.788615] env[65121]: _type = "VirtualMachine" [ 1080.788615] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1080.788901] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d95040f9-d6ee-40cc-8584-3dbdcd47d046 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.796363] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lease: (returnval){ [ 1080.796363] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5231d233-1bc5-970b-830c-cd215a0f40f5" [ 1080.796363] env[65121]: _type = "HttpNfcLease" [ 1080.796363] env[65121]: } obtained for exporting VM: (result){ [ 1080.796363] env[65121]: value = "vm-993542" [ 1080.796363] env[65121]: _type = "VirtualMachine" [ 1080.796363] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1080.796584] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the lease: (returnval){ [ 1080.796584] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5231d233-1bc5-970b-830c-cd215a0f40f5" [ 1080.796584] env[65121]: _type = "HttpNfcLease" [ 1080.796584] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1080.803967] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1080.803967] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5231d233-1bc5-970b-830c-cd215a0f40f5" [ 1080.803967] env[65121]: _type = "HttpNfcLease" [ 1080.803967] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1080.810584] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1080.810822] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1080.811033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1080.811214] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1080.811380] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1080.813680] env[65121]: INFO nova.compute.manager [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Terminating instance [ 1080.884963] env[65121]: DEBUG nova.scheduler.client.report [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1081.135261] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Releasing lock "refresh_cache-347bbb23-da89-4e6f-837c-e88bf7c57a3a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1081.135778] env[65121]: DEBUG nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Instance network_info: |[{"id": "7fe57d90-496e-44ce-bab8-1f328075e76b", "address": "fa:16:3e:c5:e6:c5", "network": {"id": "7293d35b-5f53-4ff1-bb9f-a4724377a87b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-972919019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "665a22b60ef542198ef02de9d6439807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69076131-87ac-46dd-9d5d-8d1b4ea7dec6", "external-id": "nsx-vlan-transportzone-327", "segmentation_id": 327, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fe57d90-49", "ovs_interfaceid": "7fe57d90-496e-44ce-bab8-1f328075e76b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1081.136306] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:e6:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69076131-87ac-46dd-9d5d-8d1b4ea7dec6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7fe57d90-496e-44ce-bab8-1f328075e76b', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.143872] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Creating folder: Project (665a22b60ef542198ef02de9d6439807). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1081.144225] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb9d711e-c54e-472e-9063-3939847cd7c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.154870] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107164, 'name': PowerOffVM_Task, 'duration_secs': 0.227469} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.155164] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.156016] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a927e665-8df2-4be7-a6c6-34bbc58d01d3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.160792] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Created folder: Project (665a22b60ef542198ef02de9d6439807) in parent group-v993268. [ 1081.160951] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Creating folder: Instances. Parent ref: group-v993543. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1081.161273] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a53a2e02-b6e3-4487-b5bc-44f523eab2b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.178215] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07377bb-431b-4c35-810e-f827fd14c463 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.189919] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Created folder: Instances in parent group-v993543. [ 1081.190188] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1081.190392] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1081.190613] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba9cf778-e6a6-41f2-859d-7dd07a0cda07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.217922] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1081.217922] env[65121]: value = "task-5107168" [ 1081.217922] env[65121]: _type = "Task" [ 1081.217922] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.223595] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.223909] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-defced23-03bf-4d22-9808-615e4b792928 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.232162] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1081.232162] env[65121]: value = "task-5107169" [ 1081.232162] env[65121]: _type = "Task" [ 1081.232162] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.235773] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107168, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.246764] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1081.247099] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1081.247468] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.248032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.248032] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.248285] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8ac1a43-d742-487a-806b-99ea7e048f50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.260335] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.261384] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1081.261384] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e077290-8562-4d06-8bff-24f6920420cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.267371] env[65121]: DEBUG nova.compute.manager [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Received event network-changed-7fe57d90-496e-44ce-bab8-1f328075e76b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1081.267705] env[65121]: DEBUG nova.compute.manager [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Refreshing instance network info cache due to event network-changed-7fe57d90-496e-44ce-bab8-1f328075e76b. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1081.267929] env[65121]: DEBUG oslo_concurrency.lockutils [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Acquiring lock "refresh_cache-347bbb23-da89-4e6f-837c-e88bf7c57a3a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.267973] env[65121]: DEBUG oslo_concurrency.lockutils [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Acquired lock "refresh_cache-347bbb23-da89-4e6f-837c-e88bf7c57a3a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.268208] env[65121]: DEBUG nova.network.neutron [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Refreshing network info cache for port 7fe57d90-496e-44ce-bab8-1f328075e76b {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1081.275843] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1081.275843] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52816fa3-8f88-662d-7bc4-98a0a9718e1e" [ 1081.275843] env[65121]: _type = "Task" [ 1081.275843] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.289592] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52816fa3-8f88-662d-7bc4-98a0a9718e1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.309171] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1081.309171] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5231d233-1bc5-970b-830c-cd215a0f40f5" [ 1081.309171] env[65121]: _type = "HttpNfcLease" [ 1081.309171] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1081.309570] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1081.309570] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5231d233-1bc5-970b-830c-cd215a0f40f5" [ 1081.309570] env[65121]: _type = "HttpNfcLease" [ 1081.309570] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1081.310338] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e987652-5707-407a-808e-0f6b2567d2f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.320494] env[65121]: DEBUG nova.compute.manager [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1081.320813] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1081.321274] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b700b-f905-e55e-1cbd-fbc6787029b1/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1081.321353] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b700b-f905-e55e-1cbd-fbc6787029b1/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1081.323563] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1f905b-27d9-440b-bcdc-57d955d9a7f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.384708] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.386288] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb38da2a-1ff0-46cf-8170-d81324bd4e69 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.391363] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.956s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.393711] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.656s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.393933] env[65121]: DEBUG nova.objects.instance [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lazy-loading 'resources' on Instance uuid abd87ca7-5d95-4b22-8d2a-1f37abd18f86 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.415516] env[65121]: INFO nova.scheduler.client.report [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleted allocations for instance ec463cad-4c80-4636-bc7a-9ec298a07d96 [ 1081.425657] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4914f689-c124-44f8-ac6e-115bcd515787 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.454719] env[65121]: DEBUG nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1081.459856] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.459964] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.460159] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleting the datastore file [datastore1] 97fa5791-28cb-4e77-bad7-f2cbce8bd06d {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.460434] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4534f37f-5df1-43d0-ac7a-b4170d593bea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.472800] env[65121]: DEBUG oslo_vmware.api [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1081.472800] env[65121]: value = "task-5107171" [ 1081.472800] env[65121]: _type = "Task" [ 1081.472800] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.485561] env[65121]: DEBUG oslo_vmware.api [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.488294] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1081.488549] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1081.488752] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1081.488992] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1081.489193] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1081.489376] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1081.489685] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.489872] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1081.490098] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1081.490350] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1081.490581] env[65121]: DEBUG nova.virt.hardware [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1081.492061] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0db61f-b9da-41e6-a3ab-59df9033eebf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.503462] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9523f670-66b1-4481-9b1c-08bbe3957f7e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.524618] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.534726] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Creating folder: Project (03f0dc06bd6e4448acfa12f1ae8a42f3). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1081.536016] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b40299e-205e-445a-b794-7c1931a8469f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.551755] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Created folder: Project (03f0dc06bd6e4448acfa12f1ae8a42f3) in parent group-v993268. [ 1081.551957] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Creating folder: Instances. Parent ref: group-v993546. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1081.552241] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8d5725f-9e94-46fb-9e5c-0797b0bd7a85 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.566078] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Created folder: Instances in parent group-v993546. [ 1081.566657] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1081.566657] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1081.566861] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-729bc9b1-5815-478c-9c15-32052f732958 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.590037] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1081.590037] env[65121]: value = "task-5107174" [ 1081.590037] env[65121]: _type = "Task" [ 1081.590037] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.598446] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107174, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.728076] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107168, 'name': CreateVM_Task, 'duration_secs': 0.372962} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.728268] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1081.728794] env[65121]: WARNING neutronclient.v2_0.client [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.729203] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.729352] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.729759] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1081.729997] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8fb6a88-7e23-470e-8c62-cc1d4100e5b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.735547] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1081.735547] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52221e91-ba1a-acc0-6e92-bd6567633257" [ 1081.735547] env[65121]: _type = "Task" [ 1081.735547] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.744283] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52221e91-ba1a-acc0-6e92-bd6567633257, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.772161] env[65121]: WARNING neutronclient.v2_0.client [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.772898] env[65121]: WARNING openstack [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.773260] env[65121]: WARNING openstack [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.791098] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52816fa3-8f88-662d-7bc4-98a0a9718e1e, 'name': SearchDatastore_Task, 'duration_secs': 0.012813} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.792070] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73c7235a-117e-47ce-86aa-f43c5d4c90cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.798368] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1081.798368] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f8f33a-1459-e2e2-6e99-f2f7805ad3f1" [ 1081.798368] env[65121]: _type = "Task" [ 1081.798368] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.807320] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f8f33a-1459-e2e2-6e99-f2f7805ad3f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.889304] env[65121]: WARNING openstack [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.889652] env[65121]: WARNING openstack [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.924135] env[65121]: DEBUG oslo_concurrency.lockutils [None req-db390ac8-ee90-4279-8f3b-1f5e2709eaf0 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "ec463cad-4c80-4636-bc7a-9ec298a07d96" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.542s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.962655] env[65121]: WARNING neutronclient.v2_0.client [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.963502] env[65121]: WARNING openstack [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.964017] env[65121]: WARNING openstack [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.986205] env[65121]: DEBUG oslo_vmware.api [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234094} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.989256] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.989506] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.989688] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.989902] env[65121]: INFO nova.compute.manager [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Took 0.67 seconds to destroy the instance on the hypervisor. [ 1081.990273] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1081.991118] env[65121]: DEBUG nova.compute.manager [-] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1081.991241] env[65121]: DEBUG nova.network.neutron [-] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1081.991569] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.992589] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.992589] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.040709] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.068414] env[65121]: DEBUG nova.network.neutron [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Updated VIF entry in instance network info cache for port 7fe57d90-496e-44ce-bab8-1f328075e76b. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1082.069030] env[65121]: DEBUG nova.network.neutron [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Updating instance_info_cache with network_info: [{"id": "7fe57d90-496e-44ce-bab8-1f328075e76b", "address": "fa:16:3e:c5:e6:c5", "network": {"id": "7293d35b-5f53-4ff1-bb9f-a4724377a87b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-972919019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "665a22b60ef542198ef02de9d6439807", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69076131-87ac-46dd-9d5d-8d1b4ea7dec6", "external-id": "nsx-vlan-transportzone-327", "segmentation_id": 327, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fe57d90-49", "ovs_interfaceid": "7fe57d90-496e-44ce-bab8-1f328075e76b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1082.103981] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107174, 'name': CreateVM_Task, 'duration_secs': 0.306925} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.104462] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1082.105086] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.233211] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98eb2260-a8ac-4bf4-8ba1-d88ee1b89608 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.249741] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afce6080-a1d8-49be-bf1f-5d620088fc30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.257026] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52221e91-ba1a-acc0-6e92-bd6567633257, 'name': SearchDatastore_Task, 'duration_secs': 0.010625} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.257026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.257026] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.257026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.257026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.257026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1082.257026] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8297821b-dbde-4e53-87aa-453c75f2825e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.293944] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30377dce-8a24-414d-b047-da3bec6ca479 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.299089] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1082.299089] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523c1b52-7bd4-5565-48fc-529eca5c87f6" [ 1082.299089] env[65121]: _type = "Task" [ 1082.299089] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.311984] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789124aa-13ca-42e8-92e0-b0cc54eb27dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.320763] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f8f33a-1459-e2e2-6e99-f2f7805ad3f1, 'name': SearchDatastore_Task, 'duration_secs': 0.014736} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.322093] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.322093] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. {{(pid=65121) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1082.322311] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.322501] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.322788] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c0d6fb2-18f6-49fa-a425-e2da0188db73 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.338067] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea0ccb05-ecda-4532-90e0-93792304c559 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.340832] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523c1b52-7bd4-5565-48fc-529eca5c87f6, 'name': SearchDatastore_Task, 'duration_secs': 0.012106} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.341134] env[65121]: DEBUG nova.compute.provider_tree [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.343684] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.343684] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.343684] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.353118] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1082.353118] env[65121]: value = "task-5107175" [ 1082.353118] env[65121]: _type = "Task" [ 1082.353118] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.357055] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.357055] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1082.366672] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e68da0f-03eb-436a-b57c-23adf80c87eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.380136] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1082.380136] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526c8a68-5dae-172e-1476-4cc7a8d1c2ba" [ 1082.380136] env[65121]: _type = "Task" [ 1082.380136] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.383582] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.397019] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526c8a68-5dae-172e-1476-4cc7a8d1c2ba, 'name': SearchDatastore_Task, 'duration_secs': 0.011604} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.397019] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcbdcb23-58dd-4c6b-8f2a-5afb6d78c43e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.402936] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1082.402936] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5276fec2-f89e-8c3c-14ae-8b02b705b3a6" [ 1082.402936] env[65121]: _type = "Task" [ 1082.402936] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.412491] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5276fec2-f89e-8c3c-14ae-8b02b705b3a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.495815] env[65121]: DEBUG nova.compute.manager [req-24ece3a2-b90e-48aa-99c2-627cb805ba52 req-bfc7135d-8dad-4d72-8bdf-e753a9f75ff6 service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Received event network-vif-deleted-b7cad5ed-15a4-4f85-8d54-51dcbbd9545e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1082.495815] env[65121]: INFO nova.compute.manager [req-24ece3a2-b90e-48aa-99c2-627cb805ba52 req-bfc7135d-8dad-4d72-8bdf-e753a9f75ff6 service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Neutron deleted interface b7cad5ed-15a4-4f85-8d54-51dcbbd9545e; detaching it from the instance and deleting it from the info cache [ 1082.496129] env[65121]: DEBUG nova.network.neutron [req-24ece3a2-b90e-48aa-99c2-627cb805ba52 req-bfc7135d-8dad-4d72-8bdf-e753a9f75ff6 service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1082.575684] env[65121]: DEBUG oslo_concurrency.lockutils [req-2b722fd6-06a3-4bd2-af9a-9f540732086f req-6240170a-829b-4954-abb3-63a147a3b4f1 service nova] Releasing lock "refresh_cache-347bbb23-da89-4e6f-837c-e88bf7c57a3a" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.846060] env[65121]: DEBUG nova.scheduler.client.report [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1082.866075] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498911} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.867194] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. [ 1082.868296] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438bdf88-e29a-4f96-a48d-2c6a124c0173 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.889781] env[65121]: DEBUG nova.network.neutron [-] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1082.899327] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.900433] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24146393-ce91-474c-9a88-fc78fc8dcd60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.926447] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5276fec2-f89e-8c3c-14ae-8b02b705b3a6, 'name': SearchDatastore_Task, 'duration_secs': 0.010283} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.928547] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.928761] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 347bbb23-da89-4e6f-837c-e88bf7c57a3a/347bbb23-da89-4e6f-837c-e88bf7c57a3a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1082.929114] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1082.929114] env[65121]: value = "task-5107176" [ 1082.929114] env[65121]: _type = "Task" [ 1082.929114] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.929319] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.929611] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.929964] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c41d6d7-f9d0-4fd0-bc0d-2a557d27b54b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.932243] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-008f397d-acf5-4beb-a2dd-dee58c3f6144 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.945748] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107176, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.948425] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.948640] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1082.949588] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1082.949588] env[65121]: value = "task-5107177" [ 1082.949588] env[65121]: _type = "Task" [ 1082.949588] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.949968] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00067317-9ac4-4a60-a376-59139cb49507 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.961364] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1082.961364] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5256aa1c-48e4-5ef5-bd89-bbabbc4d7bc2" [ 1082.961364] env[65121]: _type = "Task" [ 1082.961364] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.968566] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.975701] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5256aa1c-48e4-5ef5-bd89-bbabbc4d7bc2, 'name': SearchDatastore_Task, 'duration_secs': 0.012891} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.976685] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c162444-dc75-4469-90f0-a7c7e9e2ad4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.982832] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1082.982832] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b96fa1-d6e4-e8d9-d652-18a22b723b56" [ 1082.982832] env[65121]: _type = "Task" [ 1082.982832] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.991961] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b96fa1-d6e4-e8d9-d652-18a22b723b56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.998376] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfd6c349-9255-4e22-b899-1fff7b1db0a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.008631] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b22d89e-49fc-4f41-b01d-3dedf9d2e10b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.053469] env[65121]: DEBUG nova.compute.manager [req-24ece3a2-b90e-48aa-99c2-627cb805ba52 req-bfc7135d-8dad-4d72-8bdf-e753a9f75ff6 service nova] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Detach interface failed, port_id=b7cad5ed-15a4-4f85-8d54-51dcbbd9545e, reason: Instance 97fa5791-28cb-4e77-bad7-f2cbce8bd06d could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1083.352825] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.355152] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.041s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.355428] env[65121]: DEBUG nova.objects.instance [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lazy-loading 'resources' on Instance uuid 3d6fc9df-6872-4929-8813-bec1db1ebb16 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.390485] env[65121]: INFO nova.scheduler.client.report [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Deleted allocations for instance abd87ca7-5d95-4b22-8d2a-1f37abd18f86 [ 1083.400704] env[65121]: INFO nova.compute.manager [-] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Took 1.41 seconds to deallocate network for instance. [ 1083.443740] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107176, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.466054] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48638} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.466054] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 347bbb23-da89-4e6f-837c-e88bf7c57a3a/347bbb23-da89-4e6f-837c-e88bf7c57a3a.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1083.466054] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.466372] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2b044ff-cf5d-440e-bb6e-3007f7942706 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.473849] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1083.473849] env[65121]: value = "task-5107178" [ 1083.473849] env[65121]: _type = "Task" [ 1083.473849] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.483839] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.496482] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b96fa1-d6e4-e8d9-d652-18a22b723b56, 'name': SearchDatastore_Task, 'duration_secs': 0.009594} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.496783] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.497065] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1083.497398] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7dc14c3-c93f-4bfe-96cf-77813685d571 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.506542] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1083.506542] env[65121]: value = "task-5107179" [ 1083.506542] env[65121]: _type = "Task" [ 1083.506542] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.518061] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107179, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.902722] env[65121]: DEBUG oslo_concurrency.lockutils [None req-59ac51ba-daa6-43d6-aae9-def221e7a219 tempest-ListServersNegativeTestJSON-1752510963 tempest-ListServersNegativeTestJSON-1752510963-project-member] Lock "abd87ca7-5d95-4b22-8d2a-1f37abd18f86" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.889s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.909688] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.949614] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107176, 'name': ReconfigVM_Task, 'duration_secs': 0.625453} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.950317] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 03a13029-55b1-4757-80a6-9293471434dc/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.951641] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ef25aa-026c-4ce1-aa56-2eaa9893c5f1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.987676] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9460bbe3-b41a-4b29-bcde-a2d4d0c7247b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.004988] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075072} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.005309] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1084.006699] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f6bd82-234c-4bcb-88ed-c5933a770c17 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.014805] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1084.014805] env[65121]: value = "task-5107180" [ 1084.014805] env[65121]: _type = "Task" [ 1084.014805] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.042507] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 347bbb23-da89-4e6f-837c-e88bf7c57a3a/347bbb23-da89-4e6f-837c-e88bf7c57a3a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.047306] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-401ceca6-8d3e-48dc-8ea7-d1ec933ce5d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.062801] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107179, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475285} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.066222] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1084.066625] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1084.067823] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9153ca39-8113-4435-aa06-8b9fd2ee8d89 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.073522] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107180, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.075910] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1084.075910] env[65121]: value = "task-5107181" [ 1084.075910] env[65121]: _type = "Task" [ 1084.075910] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.083993] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1084.083993] env[65121]: value = "task-5107182" [ 1084.083993] env[65121]: _type = "Task" [ 1084.083993] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.096832] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107181, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.100567] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107182, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.255676] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa60b86-e87a-4c49-a991-46eec5a848e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.264062] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d798b1-aa32-476e-9eb9-c5eda5bd3401 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.302527] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef465532-48fa-4f41-b508-6daba83c7cdc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.312445] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02876bb-c64a-406d-a9a7-cd83edd63b54 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.331284] env[65121]: DEBUG nova.compute.provider_tree [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.530663] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107180, 'name': ReconfigVM_Task, 'duration_secs': 0.195806} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.531149] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.531487] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44d594b0-1561-4a68-8600-dd3edc35fc1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.541981] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1084.541981] env[65121]: value = "task-5107183" [ 1084.541981] env[65121]: _type = "Task" [ 1084.541981] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.555308] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.587860] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107181, 'name': ReconfigVM_Task, 'duration_secs': 0.332745} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.591643] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 347bbb23-da89-4e6f-837c-e88bf7c57a3a/347bbb23-da89-4e6f-837c-e88bf7c57a3a.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.592374] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92d5b799-2a67-4956-94e1-8a54db0fa4cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.600193] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107182, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081899} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.601777] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1084.602197] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1084.602197] env[65121]: value = "task-5107184" [ 1084.602197] env[65121]: _type = "Task" [ 1084.602197] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.603577] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614ed77e-455e-4d62-88d7-00fe0e4b68e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.629199] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.632280] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-779ec023-418f-4a0e-b874-d37c5e40f447 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.647605] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107184, 'name': Rename_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.653923] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1084.653923] env[65121]: value = "task-5107185" [ 1084.653923] env[65121]: _type = "Task" [ 1084.653923] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.664356] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.834904] env[65121]: DEBUG nova.scheduler.client.report [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1085.053417] env[65121]: DEBUG oslo_vmware.api [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107183, 'name': PowerOnVM_Task, 'duration_secs': 0.470296} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.053699] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.056750] env[65121]: DEBUG nova.compute.manager [None req-03812107-a070-42f5-81a0-3e03ea199ea7 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1085.057702] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dac1f8d-6118-4b66-8729-72ab2088ddb4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.120231] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107184, 'name': Rename_Task, 'duration_secs': 0.192194} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.121049] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1085.121049] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a89d9b2-6d48-4c29-91bc-91ac4e2af2fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.129865] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1085.129865] env[65121]: value = "task-5107186" [ 1085.129865] env[65121]: _type = "Task" [ 1085.129865] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.141388] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107186, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.169781] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107185, 'name': ReconfigVM_Task, 'duration_secs': 0.294907} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.170343] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.171702] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d69104d-8b4d-4092-b99f-14134710ebd7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.180510] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1085.180510] env[65121]: value = "task-5107187" [ 1085.180510] env[65121]: _type = "Task" [ 1085.180510] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.190630] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107187, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.341231] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.986s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.344566] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.685s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.346545] env[65121]: INFO nova.compute.claims [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.372255] env[65121]: INFO nova.scheduler.client.report [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted allocations for instance 3d6fc9df-6872-4929-8813-bec1db1ebb16 [ 1085.643213] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107186, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.691414] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107187, 'name': Rename_Task, 'duration_secs': 0.188152} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.691893] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1085.692193] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e862a12-60a3-4324-8536-19b5647c0f0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.700774] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1085.700774] env[65121]: value = "task-5107188" [ 1085.700774] env[65121]: _type = "Task" [ 1085.700774] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.711224] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.881786] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fefc2676-284f-47b4-a5b0-cba4751dced1 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "3d6fc9df-6872-4929-8813-bec1db1ebb16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.559s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.143487] env[65121]: DEBUG oslo_vmware.api [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107186, 'name': PowerOnVM_Task, 'duration_secs': 0.824684} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.143753] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.143968] env[65121]: INFO nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Took 7.21 seconds to spawn the instance on the hypervisor. [ 1086.144237] env[65121]: DEBUG nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1086.145064] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144c4d39-9b30-4644-a05d-fb11f4da583e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.217747] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107188, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.367341] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "efb98d37-4162-4249-9f85-008d4537db87" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.367714] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "efb98d37-4162-4249-9f85-008d4537db87" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.367922] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "efb98d37-4162-4249-9f85-008d4537db87-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.368147] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "efb98d37-4162-4249-9f85-008d4537db87-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.368339] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "efb98d37-4162-4249-9f85-008d4537db87-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.370809] env[65121]: INFO nova.compute.manager [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Terminating instance [ 1086.660737] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9d6d31-b4b3-475e-bea4-58ade429c14e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.666982] env[65121]: INFO nova.compute.manager [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Took 21.88 seconds to build instance. [ 1086.672571] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2411932-736e-4722-a76d-ce6dfbf42cea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.712482] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80e28b2-fb12-4dee-ba7e-0f1e5df562d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.724698] env[65121]: DEBUG oslo_vmware.api [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107188, 'name': PowerOnVM_Task, 'duration_secs': 0.514984} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.725587] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.725803] env[65121]: INFO nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Took 5.27 seconds to spawn the instance on the hypervisor. [ 1086.725976] env[65121]: DEBUG nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1086.727293] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c88f27d-e156-43eb-90ce-a86584249570 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.733469] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2014b7d4-fdcb-4c77-a6ab-bca627133454 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.755421] env[65121]: DEBUG nova.compute.provider_tree [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.876016] env[65121]: DEBUG nova.compute.manager [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1086.876499] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1086.877808] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5642b93a-0185-44af-a971-7506685e36a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.887901] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.888314] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3176bf94-4bbc-4965-aee5-c73378922968 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.896664] env[65121]: DEBUG oslo_vmware.api [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1086.896664] env[65121]: value = "task-5107189" [ 1086.896664] env[65121]: _type = "Task" [ 1086.896664] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.911668] env[65121]: DEBUG oslo_vmware.api [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.922833] env[65121]: DEBUG nova.compute.manager [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1086.923143] env[65121]: DEBUG nova.compute.manager [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing instance network info cache due to event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1086.923426] env[65121]: DEBUG oslo_concurrency.lockutils [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Acquiring lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.923588] env[65121]: DEBUG oslo_concurrency.lockutils [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Acquired lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.924213] env[65121]: DEBUG nova.network.neutron [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1087.170030] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bb5ec798-6579-460a-8dbb-6bf4618e241f tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.404s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1087.254880] env[65121]: INFO nova.compute.manager [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Took 19.27 seconds to build instance. [ 1087.259453] env[65121]: DEBUG nova.scheduler.client.report [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1087.411027] env[65121]: DEBUG oslo_vmware.api [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107189, 'name': PowerOffVM_Task, 'duration_secs': 0.290383} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.411027] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.411027] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1087.411027] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8fd46944-4e83-4c07-befc-7908c7663936 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.428930] env[65121]: WARNING neutronclient.v2_0.client [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.428930] env[65121]: WARNING openstack [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.429093] env[65121]: WARNING openstack [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.492200] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1087.492372] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1087.492783] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleting the datastore file [datastore1] efb98d37-4162-4249-9f85-008d4537db87 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.492912] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c3ddb98-d064-4d62-ada2-ce534e23d3e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.503134] env[65121]: DEBUG oslo_vmware.api [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1087.503134] env[65121]: value = "task-5107191" [ 1087.503134] env[65121]: _type = "Task" [ 1087.503134] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.513488] env[65121]: DEBUG oslo_vmware.api [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107191, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.566874] env[65121]: WARNING openstack [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.567300] env[65121]: WARNING openstack [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.650315] env[65121]: WARNING neutronclient.v2_0.client [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.651053] env[65121]: WARNING openstack [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.651483] env[65121]: WARNING openstack [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.758389] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a79fb837-b011-470e-83cc-c686057add41 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "39ca7a89-c262-4169-91b1-92ea45744518" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.782s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1087.770706] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1087.771356] env[65121]: DEBUG nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1087.777073] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.497s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.777655] env[65121]: INFO nova.compute.claims [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.827729] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "interface-347bbb23-da89-4e6f-837c-e88bf7c57a3a-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.828052] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "interface-347bbb23-da89-4e6f-837c-e88bf7c57a3a-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.828387] env[65121]: DEBUG nova.objects.instance [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lazy-loading 'flavor' on Instance uuid 347bbb23-da89-4e6f-837c-e88bf7c57a3a {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.015562] env[65121]: DEBUG oslo_vmware.api [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107191, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301826} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.015562] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1088.015914] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1088.015914] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1088.016159] env[65121]: INFO nova.compute.manager [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: efb98d37-4162-4249-9f85-008d4537db87] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1088.016509] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1088.016790] env[65121]: DEBUG nova.compute.manager [-] [instance: efb98d37-4162-4249-9f85-008d4537db87] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1088.017093] env[65121]: DEBUG nova.network.neutron [-] [instance: efb98d37-4162-4249-9f85-008d4537db87] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1088.017403] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.018099] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.018452] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.106672] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.167078] env[65121]: DEBUG nova.network.neutron [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updated VIF entry in instance network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1088.167530] env[65121]: DEBUG nova.network.neutron [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.283016] env[65121]: DEBUG nova.compute.utils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1088.289415] env[65121]: DEBUG nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1088.289415] env[65121]: DEBUG nova.network.neutron [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1088.289415] env[65121]: WARNING neutronclient.v2_0.client [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.289415] env[65121]: WARNING neutronclient.v2_0.client [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.290213] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.290684] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.334378] env[65121]: WARNING neutronclient.v2_0.client [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.337270] env[65121]: WARNING openstack [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.338798] env[65121]: WARNING openstack [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.351108] env[65121]: DEBUG nova.objects.instance [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lazy-loading 'pci_requests' on Instance uuid 347bbb23-da89-4e6f-837c-e88bf7c57a3a {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.475445] env[65121]: DEBUG nova.policy [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1088.671371] env[65121]: DEBUG oslo_concurrency.lockutils [req-64fa49a1-6e8b-4e75-9b7a-d65fd8583a9e req-839fd4a8-2323-41bc-82e0-5db51b1797aa service nova] Releasing lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.802128] env[65121]: DEBUG nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1088.857408] env[65121]: DEBUG nova.objects.base [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Object Instance<347bbb23-da89-4e6f-837c-e88bf7c57a3a> lazy-loaded attributes: flavor,pci_requests {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1088.857707] env[65121]: DEBUG nova.network.neutron [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1088.858063] env[65121]: WARNING neutronclient.v2_0.client [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.858409] env[65121]: WARNING neutronclient.v2_0.client [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.858975] env[65121]: WARNING openstack [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.859382] env[65121]: WARNING openstack [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.892759] env[65121]: DEBUG nova.network.neutron [-] [instance: efb98d37-4162-4249-9f85-008d4537db87] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.901732] env[65121]: DEBUG nova.network.neutron [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Successfully created port: 9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1088.999020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bf1c9009-723b-4fac-9bab-3a5248eedd3d tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "interface-347bbb23-da89-4e6f-837c-e88bf7c57a3a-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.171s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.195882] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e5b5af-5664-4070-84e1-6e4190262b1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.204951] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad050853-4e3e-4d1b-9f98-8941757d51bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.242249] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce79db98-6a9c-40b3-af62-a7c4eceb63b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.251436] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5429c703-b410-4514-9b78-e55f3cee2768 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.266934] env[65121]: DEBUG nova.compute.provider_tree [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.401313] env[65121]: INFO nova.compute.manager [-] [instance: efb98d37-4162-4249-9f85-008d4537db87] Took 1.38 seconds to deallocate network for instance. [ 1089.773032] env[65121]: DEBUG nova.scheduler.client.report [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1089.816910] env[65121]: DEBUG nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1089.843470] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1089.843851] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1089.844110] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1089.844843] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1089.844949] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1089.845096] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1089.845389] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.845573] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1089.845744] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1089.846675] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1089.846675] env[65121]: DEBUG nova.virt.hardware [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1089.847111] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c5a822-54eb-407c-b853-76577fcbc8a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.857802] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8256046-7635-4c12-a00f-8f806b596ab6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.864091] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b700b-f905-e55e-1cbd-fbc6787029b1/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1089.864903] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5962f5b8-7923-4373-b21a-3efcfeb83d73 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.880710] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b700b-f905-e55e-1cbd-fbc6787029b1/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1089.880884] env[65121]: ERROR oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b700b-f905-e55e-1cbd-fbc6787029b1/disk-0.vmdk due to incomplete transfer. [ 1089.881141] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-261bc631-7089-4b76-baae-0fa255e5da9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.890890] env[65121]: DEBUG oslo_vmware.rw_handles [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b700b-f905-e55e-1cbd-fbc6787029b1/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1089.891120] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Uploaded image 872a03ef-a1b4-487f-afd6-17a5c0d28e41 to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1089.893725] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1089.894030] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fefe45cd-f49f-4e7c-bfab-204fcddc4705 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.902491] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1089.902491] env[65121]: value = "task-5107192" [ 1089.902491] env[65121]: _type = "Task" [ 1089.902491] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.914811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.915197] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107192, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.278492] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.279205] env[65121]: DEBUG nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1090.283910] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.297s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.416229] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107192, 'name': Destroy_Task, 'duration_secs': 0.360214} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.416581] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Destroyed the VM [ 1090.416840] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1090.417123] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4a5a8700-690c-47d5-830d-858a1200badc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.425899] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1090.425899] env[65121]: value = "task-5107193" [ 1090.425899] env[65121]: _type = "Task" [ 1090.425899] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.438336] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107193, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.505608] env[65121]: DEBUG nova.compute.manager [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1090.506271] env[65121]: DEBUG nova.compute.manager [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing instance network info cache due to event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1090.506974] env[65121]: DEBUG oslo_concurrency.lockutils [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Acquiring lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.507173] env[65121]: DEBUG oslo_concurrency.lockutils [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Acquired lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1090.507402] env[65121]: DEBUG nova.network.neutron [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1090.603848] env[65121]: DEBUG nova.network.neutron [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Successfully updated port: 9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1090.788052] env[65121]: DEBUG nova.compute.utils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1090.795540] env[65121]: DEBUG nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1090.795540] env[65121]: DEBUG nova.network.neutron [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1090.795777] env[65121]: WARNING neutronclient.v2_0.client [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1090.800021] env[65121]: WARNING neutronclient.v2_0.client [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1090.800021] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1090.800021] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1090.876025] env[65121]: DEBUG nova.policy [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '175d752afbc745258500be8624bfbf14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e491b88dcaad42dcb345b3217bc6f941', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1090.943499] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107193, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.989818] env[65121]: DEBUG nova.compute.manager [req-7e9127dc-8225-4afe-b2f2-5524db22e71b req-6ea90499-bebb-4e22-b85e-dd8a7ef4966d service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-vif-plugged-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1090.991465] env[65121]: DEBUG oslo_concurrency.lockutils [req-7e9127dc-8225-4afe-b2f2-5524db22e71b req-6ea90499-bebb-4e22-b85e-dd8a7ef4966d service nova] Acquiring lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.991465] env[65121]: DEBUG oslo_concurrency.lockutils [req-7e9127dc-8225-4afe-b2f2-5524db22e71b req-6ea90499-bebb-4e22-b85e-dd8a7ef4966d service nova] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.991465] env[65121]: DEBUG oslo_concurrency.lockutils [req-7e9127dc-8225-4afe-b2f2-5524db22e71b req-6ea90499-bebb-4e22-b85e-dd8a7ef4966d service nova] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.991465] env[65121]: DEBUG nova.compute.manager [req-7e9127dc-8225-4afe-b2f2-5524db22e71b req-6ea90499-bebb-4e22-b85e-dd8a7ef4966d service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] No waiting events found dispatching network-vif-plugged-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1090.991465] env[65121]: WARNING nova.compute.manager [req-7e9127dc-8225-4afe-b2f2-5524db22e71b req-6ea90499-bebb-4e22-b85e-dd8a7ef4966d service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received unexpected event network-vif-plugged-9dc8b4e9-1442-43ba-a663-567f89e3a428 for instance with vm_state building and task_state spawning. [ 1091.012499] env[65121]: WARNING neutronclient.v2_0.client [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.014367] env[65121]: WARNING openstack [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.014736] env[65121]: WARNING openstack [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.109498] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.109498] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.109498] env[65121]: DEBUG nova.network.neutron [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1091.145711] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d66df8-722f-4103-8db0-f008a11c9dfe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.153887] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b5f295-b446-4152-bf77-f9c60c63f225 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.193427] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47457c94-c4d8-4954-8585-a55988a4d7ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.207896] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171f9fb8-456d-404a-bcf9-24077bfa9d8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.224967] env[65121]: DEBUG nova.compute.provider_tree [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.268487] env[65121]: DEBUG nova.network.neutron [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Successfully created port: b3bafe6b-71bc-45d0-bf37-29da60b15de0 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1091.295852] env[65121]: DEBUG nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1091.443169] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107193, 'name': RemoveSnapshot_Task, 'duration_secs': 0.526423} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.443744] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1091.444166] env[65121]: DEBUG nova.compute.manager [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1091.445345] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab062960-9511-4a38-bd6f-cb4ab5f05653 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.455772] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.456175] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.456461] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.456718] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.456970] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.461042] env[65121]: INFO nova.compute.manager [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Terminating instance [ 1091.576263] env[65121]: WARNING openstack [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.577026] env[65121]: WARNING openstack [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.613702] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.614111] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.706697] env[65121]: DEBUG nova.network.neutron [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1091.728629] env[65121]: DEBUG nova.scheduler.client.report [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1091.747458] env[65121]: WARNING neutronclient.v2_0.client [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1091.749062] env[65121]: WARNING openstack [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.749062] env[65121]: WARNING openstack [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.762005] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1091.762401] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1091.963473] env[65121]: INFO nova.compute.manager [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Shelve offloading [ 1091.971105] env[65121]: DEBUG nova.compute.manager [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1091.971105] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.971653] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc9c9b9-a6b8-4566-8317-a317d6fa9858 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.981625] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.982665] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8917ddc-3e36-4c73-a586-d036e81c5859 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.991442] env[65121]: DEBUG oslo_vmware.api [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1091.991442] env[65121]: value = "task-5107194" [ 1091.991442] env[65121]: _type = "Task" [ 1091.991442] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.002730] env[65121]: DEBUG oslo_vmware.api [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.251545] env[65121]: DEBUG nova.network.neutron [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updated VIF entry in instance network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1092.251943] env[65121]: DEBUG nova.network.neutron [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.280041] env[65121]: WARNING neutronclient.v2_0.client [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.281917] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.281917] env[65121]: WARNING openstack [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.316503] env[65121]: DEBUG nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1092.361167] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1092.361517] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1092.361756] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1092.361981] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1092.362206] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1092.362399] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1092.362700] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1092.362932] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1092.363860] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1092.363860] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1092.363860] env[65121]: DEBUG nova.virt.hardware [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1092.365346] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a6114a-0867-461e-a5b0-78387c8d1f4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.378358] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d88a698-c103-4cd2-8ef7-65637a2ff81c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.470293] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.470655] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a68435a-727b-4be3-95b7-5c1d3cf1a6f6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.478739] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1092.478739] env[65121]: value = "task-5107198" [ 1092.478739] env[65121]: _type = "Task" [ 1092.478739] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.485963] env[65121]: DEBUG nova.network.neutron [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.496693] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1092.497477] env[65121]: DEBUG nova.compute.manager [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1092.498363] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf316419-f8d8-494e-a93c-635f0b99bdca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.508322] env[65121]: DEBUG oslo_vmware.api [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.510597] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.510819] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1092.511114] env[65121]: DEBUG nova.network.neutron [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1092.740272] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.456s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.746637] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.449s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.746637] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.749075] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.839s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.749360] env[65121]: DEBUG nova.objects.instance [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'resources' on Instance uuid 97fa5791-28cb-4e77-bad7-f2cbce8bd06d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.757295] env[65121]: DEBUG oslo_concurrency.lockutils [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] Releasing lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.758349] env[65121]: DEBUG nova.compute.manager [req-018d52ed-e2f8-490a-bafd-6c16246950f1 req-d70c0918-27c4-4471-99be-3422bad8e3fe service nova] [instance: efb98d37-4162-4249-9f85-008d4537db87] Received event network-vif-deleted-adb6945c-5a1e-487c-b116-4200874d80a0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1092.783547] env[65121]: INFO nova.scheduler.client.report [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted allocations for instance 564159fe-6ecd-4276-8d65-cbe25859493a [ 1092.993642] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.994120] env[65121]: DEBUG nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Instance network_info: |[{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1092.998634] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:56:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9ec24851-7bb6-426b-b28f-f7b246df1713', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9dc8b4e9-1442-43ba-a663-567f89e3a428', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1093.006615] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1093.006875] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1093.007623] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64013435-2cec-420e-a1de-5309d132b26b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.026954] env[65121]: WARNING neutronclient.v2_0.client [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.027663] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.028026] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.035316] env[65121]: DEBUG oslo_vmware.api [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107194, 'name': PowerOffVM_Task, 'duration_secs': 0.819416} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.036226] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.036401] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.036719] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1ee3e37-c38d-4e73-a176-0ee2bfbd0a3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.042271] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1093.042271] env[65121]: value = "task-5107199" [ 1093.042271] env[65121]: _type = "Task" [ 1093.042271] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.052916] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107199, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.109542] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1093.109824] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1093.110061] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Deleting the datastore file [datastore1] 347bbb23-da89-4e6f-837c-e88bf7c57a3a {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.111191] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9aec4e8d-e79e-4508-a25c-696c30e2ca0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.122650] env[65121]: DEBUG oslo_vmware.api [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for the task: (returnval){ [ 1093.122650] env[65121]: value = "task-5107201" [ 1093.122650] env[65121]: _type = "Task" [ 1093.122650] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.136655] env[65121]: DEBUG oslo_vmware.api [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107201, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.184956] env[65121]: DEBUG nova.network.neutron [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Successfully updated port: b3bafe6b-71bc-45d0-bf37-29da60b15de0 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1093.190782] env[65121]: DEBUG nova.compute.manager [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1093.191656] env[65121]: DEBUG nova.compute.manager [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing instance network info cache due to event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1093.191656] env[65121]: DEBUG oslo_concurrency.lockutils [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Acquiring lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.191656] env[65121]: DEBUG oslo_concurrency.lockutils [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Acquired lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.191656] env[65121]: DEBUG nova.network.neutron [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1093.238123] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.238645] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.296836] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c286654f-93dd-4545-90c8-c05a9205a5ca tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "564159fe-6ecd-4276-8d65-cbe25859493a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.606s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.331154] env[65121]: INFO nova.scheduler.client.report [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted allocation for migration 06135b5c-3572-417f-99a9-adefc4648d55 [ 1093.392986] env[65121]: WARNING neutronclient.v2_0.client [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.394090] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.394449] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.498758] env[65121]: DEBUG nova.compute.manager [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1093.498758] env[65121]: DEBUG nova.compute.manager [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing instance network info cache due to event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1093.498758] env[65121]: DEBUG oslo_concurrency.lockutils [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.498758] env[65121]: DEBUG oslo_concurrency.lockutils [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.498758] env[65121]: DEBUG nova.network.neutron [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1093.509108] env[65121]: DEBUG nova.network.neutron [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1093.555927] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107199, 'name': CreateVM_Task, 'duration_secs': 0.356492} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.556120] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1093.556642] env[65121]: WARNING neutronclient.v2_0.client [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.557039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.557943] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.557943] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1093.558243] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca6b0360-ec4b-4317-ba6d-dafe8d802480 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.566740] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1093.566740] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52575203-8c7e-e52b-3c50-c44c0c883228" [ 1093.566740] env[65121]: _type = "Task" [ 1093.566740] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.576346] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52575203-8c7e-e52b-3c50-c44c0c883228, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.579044] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70390387-d97d-4fc5-ad2d-8373462b4f7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.586361] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adf2643-204b-4438-be36-be12570f9c20 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.617696] env[65121]: INFO nova.compute.manager [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Rebuilding instance [ 1093.620974] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eadcd3c-8a0c-4bf3-a1dc-3a01bde3be61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.629869] env[65121]: DEBUG oslo_concurrency.lockutils [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.630012] env[65121]: DEBUG oslo_concurrency.lockutils [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.636460] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a999e89-adb4-4505-b50f-d69fbabbe499 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.646145] env[65121]: DEBUG oslo_vmware.api [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Task: {'id': task-5107201, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182308} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.647533] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.648748] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1093.648748] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1093.648748] env[65121]: INFO nova.compute.manager [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Took 1.68 seconds to destroy the instance on the hypervisor. [ 1093.648748] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1093.649076] env[65121]: DEBUG nova.compute.manager [-] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1093.649076] env[65121]: DEBUG nova.network.neutron [-] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1093.649259] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.649969] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.652058] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.668862] env[65121]: DEBUG nova.compute.provider_tree [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.684439] env[65121]: DEBUG nova.compute.manager [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1093.685696] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74dbbc6-c0f6-42b6-8014-a841804f8f12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.691139] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.691445] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.691556] env[65121]: DEBUG nova.network.neutron [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1093.693362] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.695726] env[65121]: WARNING neutronclient.v2_0.client [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.696375] env[65121]: WARNING openstack [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.696710] env[65121]: WARNING openstack [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.827874] env[65121]: WARNING openstack [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.828317] env[65121]: WARNING openstack [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.839509] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3c2a621a-0fbf-4d8c-b4d1-5bf50659a15c tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 17.628s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.913188] env[65121]: WARNING neutronclient.v2_0.client [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.913751] env[65121]: WARNING openstack [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.914104] env[65121]: WARNING openstack [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.997362] env[65121]: DEBUG nova.network.neutron [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updated VIF entry in instance network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1093.997790] env[65121]: DEBUG nova.network.neutron [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.002148] env[65121]: WARNING neutronclient.v2_0.client [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.003035] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.003131] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.011801] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.012180] env[65121]: WARNING neutronclient.v2_0.client [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.012750] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.013094] env[65121]: WARNING openstack [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.018514] env[65121]: WARNING neutronclient.v2_0.client [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.081802] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52575203-8c7e-e52b-3c50-c44c0c883228, 'name': SearchDatastore_Task, 'duration_secs': 0.009925} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.082113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.082344] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1094.082576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.082752] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.082886] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1094.083184] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e90ef45-7f6f-4bf2-a278-6345e42e9443 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.100523] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1094.100703] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1094.101481] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a092d85e-f6ed-4115-9a9d-9099ae77cbdc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.110135] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1094.110135] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52f28b2d-9a35-968f-0bc9-b1cac32922e1" [ 1094.110135] env[65121]: _type = "Task" [ 1094.110135] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.115681] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.116060] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.126437] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f28b2d-9a35-968f-0bc9-b1cac32922e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.135862] env[65121]: DEBUG nova.compute.utils [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1094.171893] env[65121]: DEBUG nova.scheduler.client.report [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1094.203542] env[65121]: WARNING neutronclient.v2_0.client [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.204387] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.204737] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.213605] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.214033] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.328152] env[65121]: DEBUG nova.network.neutron [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1094.391204] env[65121]: DEBUG nova.network.neutron [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updated VIF entry in instance network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1094.392042] env[65121]: DEBUG nova.network.neutron [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.404932] env[65121]: DEBUG nova.network.neutron [-] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.419040] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.419040] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.432821] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1094.433264] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f436b4c-75a8-4b83-97a0-7b8b1edf44a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.443293] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1094.443982] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b950d3d4-03b8-4b18-9bec-5452574367b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.490920] env[65121]: WARNING neutronclient.v2_0.client [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.491761] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.492187] env[65121]: WARNING openstack [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.501356] env[65121]: DEBUG oslo_concurrency.lockutils [req-f4417951-16ad-4374-95f7-f81631dfbce1 req-abe3af38-02ab-4903-9c77-0b718ecd5d70 service nova] Releasing lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.544454] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1094.544727] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1094.546190] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleting the datastore file [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.546190] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b79306d1-b568-4387-a753-a091de104e2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.555337] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1094.555337] env[65121]: value = "task-5107203" [ 1094.555337] env[65121]: _type = "Task" [ 1094.555337] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.572234] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.606923] env[65121]: DEBUG nova.network.neutron [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance_info_cache with network_info: [{"id": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "address": "fa:16:3e:14:67:48", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3bafe6b-71", "ovs_interfaceid": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.619814] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52f28b2d-9a35-968f-0bc9-b1cac32922e1, 'name': SearchDatastore_Task, 'duration_secs': 0.03854} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.621611] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18348177-dd0e-4307-9bb7-fc5a53c97f77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.628217] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1094.628217] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52b9938e-10b3-94f7-08b7-dff9527a6aeb" [ 1094.628217] env[65121]: _type = "Task" [ 1094.628217] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.637554] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b9938e-10b3-94f7-08b7-dff9527a6aeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.639322] env[65121]: DEBUG oslo_concurrency.lockutils [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.677215] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.928s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.679913] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.765s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.680203] env[65121]: DEBUG nova.objects.instance [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lazy-loading 'resources' on Instance uuid efb98d37-4162-4249-9f85-008d4537db87 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.705544] env[65121]: INFO nova.scheduler.client.report [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted allocations for instance 97fa5791-28cb-4e77-bad7-f2cbce8bd06d [ 1094.726724] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1094.727357] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26a22243-2bd1-4510-8503-fab169c7b4e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.736428] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1094.736428] env[65121]: value = "task-5107204" [ 1094.736428] env[65121]: _type = "Task" [ 1094.736428] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.746233] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.874754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.875078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.896765] env[65121]: DEBUG oslo_concurrency.lockutils [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.898080] env[65121]: DEBUG nova.compute.manager [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1094.898080] env[65121]: DEBUG nova.compute.manager [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing instance network info cache due to event network-changed-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1094.898080] env[65121]: DEBUG oslo_concurrency.lockutils [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Acquiring lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.898228] env[65121]: DEBUG oslo_concurrency.lockutils [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Acquired lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.898455] env[65121]: DEBUG nova.network.neutron [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Refreshing network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1094.909351] env[65121]: INFO nova.compute.manager [-] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Took 1.26 seconds to deallocate network for instance. [ 1094.973834] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "03a13029-55b1-4757-80a6-9293471434dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.974396] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "03a13029-55b1-4757-80a6-9293471434dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.974878] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "03a13029-55b1-4757-80a6-9293471434dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.975058] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "03a13029-55b1-4757-80a6-9293471434dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.975324] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "03a13029-55b1-4757-80a6-9293471434dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.980159] env[65121]: INFO nova.compute.manager [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Terminating instance [ 1095.071230] env[65121]: DEBUG oslo_vmware.api [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174764} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.071230] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.071230] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1095.071230] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1095.097920] env[65121]: INFO nova.scheduler.client.report [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleted allocations for instance 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 [ 1095.110060] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.110493] env[65121]: DEBUG nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Instance network_info: |[{"id": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "address": "fa:16:3e:14:67:48", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3bafe6b-71", "ovs_interfaceid": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1095.110984] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:67:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3bafe6b-71bc-45d0-bf37-29da60b15de0', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1095.119733] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1095.120017] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1095.120260] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db6ba1e3-b29c-4c7d-909d-33dda69dcbe9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.148920] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52b9938e-10b3-94f7-08b7-dff9527a6aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.009861} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.150851] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.151202] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7/3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1095.151857] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1095.151857] env[65121]: value = "task-5107205" [ 1095.151857] env[65121]: _type = "Task" [ 1095.151857] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.151857] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5c944bd-11ce-455f-a82d-820cb1b8fa22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.167804] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107205, 'name': CreateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.168608] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1095.168608] env[65121]: value = "task-5107206" [ 1095.168608] env[65121]: _type = "Task" [ 1095.168608] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.182572] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.214559] env[65121]: DEBUG oslo_concurrency.lockutils [None req-47565445-d24f-42b5-8e59-66216e6826ee tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "97fa5791-28cb-4e77-bad7-f2cbce8bd06d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.404s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1095.249646] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107204, 'name': PowerOffVM_Task, 'duration_secs': 0.134645} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.249953] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1095.250638] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.251519] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbca8e7-1f31-4c0d-90ba-c627eceb75a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.263149] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1095.263537] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4d1d85b-f1ad-42a3-bd88-339063cc81e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.297521] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1095.297822] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1095.298146] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Deleting the datastore file [datastore1] 39ca7a89-c262-4169-91b1-92ea45744518 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.301974] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06516a63-5348-45d5-af0c-f5f81530a314 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.313068] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1095.313068] env[65121]: value = "task-5107209" [ 1095.313068] env[65121]: _type = "Task" [ 1095.313068] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.327180] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107209, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.378649] env[65121]: INFO nova.compute.manager [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Detaching volume 79bb9f39-f959-4b03-8058-0e7401b28853 [ 1095.401946] env[65121]: WARNING neutronclient.v2_0.client [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.402929] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.402929] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.418294] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.428509] env[65121]: INFO nova.virt.block_device [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Attempting to driver detach volume 79bb9f39-f959-4b03-8058-0e7401b28853 from mountpoint /dev/sdb [ 1095.428808] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1095.429044] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993520', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'name': 'volume-79bb9f39-f959-4b03-8058-0e7401b28853', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6fe86deb-0903-4769-a05c-b7d0acec0103', 'attached_at': '', 'detached_at': '', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'serial': '79bb9f39-f959-4b03-8058-0e7401b28853'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1095.430094] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e688b0c-226c-4957-bb2a-7d2a63f07744 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.473959] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c1e3b3-116e-43b8-beae-2d0c0620e1c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.479079] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Received event network-vif-plugged-b3bafe6b-71bc-45d0-bf37-29da60b15de0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1095.479771] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Acquiring lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.479771] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.479771] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1095.479941] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] No waiting events found dispatching network-vif-plugged-b3bafe6b-71bc-45d0-bf37-29da60b15de0 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1095.480078] env[65121]: WARNING nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Received unexpected event network-vif-plugged-b3bafe6b-71bc-45d0-bf37-29da60b15de0 for instance with vm_state building and task_state spawning. [ 1095.480286] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Received event network-changed-b3bafe6b-71bc-45d0-bf37-29da60b15de0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1095.480450] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Refreshing instance network info cache due to event network-changed-b3bafe6b-71bc-45d0-bf37-29da60b15de0. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1095.480604] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Acquiring lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.480792] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Acquired lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.481048] env[65121]: DEBUG nova.network.neutron [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Refreshing network info cache for port b3bafe6b-71bc-45d0-bf37-29da60b15de0 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1095.483934] env[65121]: DEBUG nova.compute.manager [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1095.484268] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.485532] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b11f82-c38f-45e2-82ad-bc27bf74b3fb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.493195] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc66bd81-077b-43f6-9e90-3277b36a7411 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.503127] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.504137] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-721f43d6-5f79-4a03-a117-14fb136796a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.544744] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839816ac-7844-4eb6-8f0c-20e121e53550 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.548898] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1095.548898] env[65121]: value = "task-5107210" [ 1095.548898] env[65121]: _type = "Task" [ 1095.548898] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.565333] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The volume has not been displaced from its original location: [datastore1] volume-79bb9f39-f959-4b03-8058-0e7401b28853/volume-79bb9f39-f959-4b03-8058-0e7401b28853.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1095.571981] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Reconfiguring VM instance instance-0000004f to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1095.573881] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.574312] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.582711] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f38a0c24-5c7c-4082-b789-d2dc4e319b2a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.606097] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.606294] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.610037] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db83132-de72-4f37-b226-ac6f4c70c7b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.621765] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25558ce6-8db2-44e1-867f-6406a3c99dd1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.625914] env[65121]: DEBUG oslo_vmware.api [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1095.625914] env[65121]: value = "task-5107211" [ 1095.625914] env[65121]: _type = "Task" [ 1095.625914] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.669160] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00382344-35b9-4ece-85d9-09a3d9fc351b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.676216] env[65121]: DEBUG oslo_vmware.api [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107211, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.685013] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107205, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.689886] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3d57db-0ab6-4b7e-a534-d80de4c11ba2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.695634] env[65121]: WARNING neutronclient.v2_0.client [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.696443] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.696911] env[65121]: WARNING openstack [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.712357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.712672] env[65121]: DEBUG oslo_concurrency.lockutils [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.712943] env[65121]: INFO nova.compute.manager [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Attaching volume 41abb581-3841-4c0e-b79e-9820e1b241f1 to /dev/sdb [ 1095.714752] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107206, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.726912] env[65121]: DEBUG nova.compute.provider_tree [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.770548] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de817438-e9e1-4522-9876-82c86d92c9ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.785039] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c790e3-edc7-4ad6-af62-3a47c42ae268 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.801834] env[65121]: DEBUG nova.virt.block_device [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updating existing volume attachment record: a4501830-18f8-4e71-838f-7ac1eec1ccf0 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1095.824136] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107209, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.407397} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.824417] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.824603] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1095.824872] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1095.843990] env[65121]: DEBUG nova.network.neutron [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updated VIF entry in instance network info cache for port 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1095.844156] env[65121]: DEBUG nova.network.neutron [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [{"id": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "address": "fa:16:3e:cf:30:98", "network": {"id": "be409d14-f237-414a-be81-f731892d2a89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1110960302-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a98eae6244314b09865c29eff2ee56a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82e421ec-a3", "ovs_interfaceid": "82e421ec-a3e6-4fcb-a1f4-e88c302e1df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1095.985316] env[65121]: WARNING neutronclient.v2_0.client [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.986109] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.986427] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.060888] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107210, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.112390] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.112644] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.112818] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.112986] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.113161] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.115246] env[65121]: INFO nova.compute.manager [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Terminating instance [ 1096.137700] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.137872] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.149095] env[65121]: DEBUG oslo_vmware.api [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107211, 'name': ReconfigVM_Task, 'duration_secs': 0.302097} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.149393] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Reconfigured VM instance instance-0000004f to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1096.155204] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74491bfb-9eeb-4977-b40f-cff209a1f93d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.191060] env[65121]: DEBUG oslo_vmware.api [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1096.191060] env[65121]: value = "task-5107215" [ 1096.191060] env[65121]: _type = "Task" [ 1096.191060] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.191380] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107205, 'name': CreateVM_Task, 'duration_secs': 0.794168} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.192824] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1096.201467] env[65121]: WARNING neutronclient.v2_0.client [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.202015] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.202205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.202639] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1096.207596] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67321f65-54a3-4675-aa44-2181af4e2c98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.210195] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107206, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693314} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.211132] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7/3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1096.211463] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.212429] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-263a6ae7-ed0c-455e-ba2e-c4d48f267ce9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.223214] env[65121]: DEBUG oslo_vmware.api [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107215, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.223214] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1096.223214] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5257b5a4-b6bf-0ac9-a35a-d54b8b78435c" [ 1096.223214] env[65121]: _type = "Task" [ 1096.223214] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.230058] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1096.230058] env[65121]: value = "task-5107216" [ 1096.230058] env[65121]: _type = "Task" [ 1096.230058] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.239376] env[65121]: DEBUG nova.scheduler.client.report [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.243235] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5257b5a4-b6bf-0ac9-a35a-d54b8b78435c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.244656] env[65121]: WARNING neutronclient.v2_0.client [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.245389] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.245761] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.261076] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107216, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.346742] env[65121]: DEBUG oslo_concurrency.lockutils [req-8babc509-3c72-4435-9004-321135aad4a9 req-7c6bb368-2988-4486-8807-9cc552c14798 service nova] Releasing lock "refresh_cache-03a13029-55b1-4757-80a6-9293471434dc" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.375702] env[65121]: DEBUG nova.network.neutron [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updated VIF entry in instance network info cache for port b3bafe6b-71bc-45d0-bf37-29da60b15de0. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1096.376104] env[65121]: DEBUG nova.network.neutron [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance_info_cache with network_info: [{"id": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "address": "fa:16:3e:14:67:48", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3bafe6b-71", "ovs_interfaceid": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1096.561377] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107210, 'name': PowerOffVM_Task, 'duration_secs': 0.596455} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.561642] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1096.561820] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1096.562109] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6cddc2ad-da86-441b-a1dc-16cfdf49abd7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.607628] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.607628] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.620276] env[65121]: DEBUG nova.compute.manager [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1096.620276] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1096.620784] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d50a7c-730a-4475-9be1-7d8297c8f414 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.631112] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1096.631112] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37c94d0e-f9bd-4191-bc1f-ffaad414c547 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.638801] env[65121]: DEBUG oslo_vmware.api [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1096.638801] env[65121]: value = "task-5107218" [ 1096.638801] env[65121]: _type = "Task" [ 1096.638801] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.651455] env[65121]: DEBUG oslo_vmware.api [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.652928] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1096.653167] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1096.653337] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Deleting the datastore file [datastore1] 03a13029-55b1-4757-80a6-9293471434dc {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.653613] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2516eb34-bd03-43b9-9a33-a6ad429a6cda {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.661075] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for the task: (returnval){ [ 1096.661075] env[65121]: value = "task-5107219" [ 1096.661075] env[65121]: _type = "Task" [ 1096.661075] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.673214] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.703360] env[65121]: DEBUG oslo_vmware.api [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107215, 'name': ReconfigVM_Task, 'duration_secs': 0.173821} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.703720] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993520', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'name': 'volume-79bb9f39-f959-4b03-8058-0e7401b28853', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '6fe86deb-0903-4769-a05c-b7d0acec0103', 'attached_at': '', 'detached_at': '', 'volume_id': '79bb9f39-f959-4b03-8058-0e7401b28853', 'serial': '79bb9f39-f959-4b03-8058-0e7401b28853'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1096.735862] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5257b5a4-b6bf-0ac9-a35a-d54b8b78435c, 'name': SearchDatastore_Task, 'duration_secs': 0.029084} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.739464] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.739752] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1096.740022] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.740382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.740382] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1096.741137] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3baf218-de68-4132-a0d7-8c74a95d0c96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.749644] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107216, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11286} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.749940] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1096.750777] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3767406e-70d0-4de8-92e1-e684b7e31ae9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.754985] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1096.755190] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1096.757300] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.077s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.759998] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9177edbe-6443-4731-8980-3dc667cc7702 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.773894] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.356s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.774286] env[65121]: DEBUG nova.objects.instance [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lazy-loading 'resources' on Instance uuid 347bbb23-da89-4e6f-837c-e88bf7c57a3a {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.784785] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7/3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.786525] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d05764a-870b-40ad-8ab9-de4e1d7888c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.806176] env[65121]: INFO nova.scheduler.client.report [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted allocations for instance efb98d37-4162-4249-9f85-008d4537db87 [ 1096.807688] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1096.807688] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52af8a9f-c720-e8f4-762e-029648afdc01" [ 1096.807688] env[65121]: _type = "Task" [ 1096.807688] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.817025] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1096.817025] env[65121]: value = "task-5107220" [ 1096.817025] env[65121]: _type = "Task" [ 1096.817025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.826938] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52af8a9f-c720-e8f4-762e-029648afdc01, 'name': SearchDatastore_Task, 'duration_secs': 0.024286} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.828274] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82502c7b-2924-4fa8-8235-58d1c84d6262 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.834450] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107220, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.841345] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1096.841345] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d1f114-4e15-6d3c-0ae5-2f94992e4bfa" [ 1096.841345] env[65121]: _type = "Task" [ 1096.841345] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.854305] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d1f114-4e15-6d3c-0ae5-2f94992e4bfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.870037] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1096.870323] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1096.870490] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1096.870681] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1096.870831] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1096.870972] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1096.871216] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1096.871372] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1096.871545] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1096.871705] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1096.871883] env[65121]: DEBUG nova.virt.hardware [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1096.873445] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5cfa08-0231-4312-83f5-7d483c3496aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.879011] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Releasing lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.879299] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Received event network-vif-deleted-7fe57d90-496e-44ce-bab8-1f328075e76b {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1096.879486] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received event network-vif-unplugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1096.879686] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.879913] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.880116] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.880267] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] No waiting events found dispatching network-vif-unplugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1096.880453] env[65121]: WARNING nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received unexpected event network-vif-unplugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca for instance with vm_state shelved_offloaded and task_state None. [ 1096.880618] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received event network-changed-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1096.880764] env[65121]: DEBUG nova.compute.manager [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Refreshing instance network info cache due to event network-changed-04d47dd2-7f48-48e4-aed4-9d693168b8ca. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1096.880958] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Acquiring lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.881115] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Acquired lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.881269] env[65121]: DEBUG nova.network.neutron [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Refreshing network info cache for port 04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1096.886603] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee30b59-ab68-454e-ab8a-bf647b4aef4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.901905] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Instance VIF info [] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1096.907539] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1096.908058] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1096.908288] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94c4c125-9db2-427b-9fb3-5cc0b950139b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.925697] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1096.925697] env[65121]: value = "task-5107221" [ 1096.925697] env[65121]: _type = "Task" [ 1096.925697] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.934284] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107221, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.109614] env[65121]: DEBUG nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1097.149920] env[65121]: DEBUG oslo_vmware.api [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107218, 'name': PowerOffVM_Task, 'duration_secs': 0.24443} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.151985] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.152201] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.155957] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72d748b2-1005-4e1c-a98c-12f0c0f55a22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.171486] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.262128] env[65121]: DEBUG nova.objects.instance [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.263696] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.263909] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.264417] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleting the datastore file [datastore1] c4096314-270a-4270-9e1d-5ace8ddbd286 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.266216] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bed89202-158a-4919-a94e-3adab0d195e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.279150] env[65121]: DEBUG oslo_vmware.api [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1097.279150] env[65121]: value = "task-5107224" [ 1097.279150] env[65121]: _type = "Task" [ 1097.279150] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.294426] env[65121]: DEBUG oslo_vmware.api [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107224, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.324344] env[65121]: DEBUG oslo_concurrency.lockutils [None req-489a4202-7103-4ad0-a0d2-64da41d4ae26 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "efb98d37-4162-4249-9f85-008d4537db87" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.956s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.332176] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107220, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.357927] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d1f114-4e15-6d3c-0ae5-2f94992e4bfa, 'name': SearchDatastore_Task, 'duration_secs': 0.017871} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.361194] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1097.361512] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 9f9118f1-c708-469f-839e-b3c4ef52f38f/9f9118f1-c708-469f-839e-b3c4ef52f38f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1097.362436] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03776ad4-2077-4a58-ba99-abcb9d4cfc48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.372761] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1097.372761] env[65121]: value = "task-5107225" [ 1097.372761] env[65121]: _type = "Task" [ 1097.372761] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.382243] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107225, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.386877] env[65121]: WARNING neutronclient.v2_0.client [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.387606] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.388056] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.436505] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107221, 'name': CreateVM_Task, 'duration_secs': 0.332962} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.439277] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1097.441674] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.441855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.442201] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1097.442721] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a552008e-4698-44d2-9fd1-fdf70a954138 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.448104] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1097.448104] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5225b99d-613f-b7fc-3a24-77845724098f" [ 1097.448104] env[65121]: _type = "Task" [ 1097.448104] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.460819] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5225b99d-613f-b7fc-3a24-77845724098f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.504645] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.505036] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.568516] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20df33c3-0a0d-4106-8306-6149103793bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.576810] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e34a68-7f35-4d73-ac38-e094ff77a945 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.613096] env[65121]: WARNING neutronclient.v2_0.client [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.613854] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.614218] env[65121]: WARNING openstack [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.623050] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d4e634-bf93-4561-b358-b7e4c5a8dde7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.640073] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668e7767-bc21-4cad-b0d6-04a1dae1f150 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.649998] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.661435] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.662188] env[65121]: DEBUG nova.compute.provider_tree [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.685191] env[65121]: DEBUG oslo_vmware.api [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Task: {'id': task-5107219, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.561025} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.685551] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.685745] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1097.685988] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1097.686204] env[65121]: INFO nova.compute.manager [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Took 2.20 seconds to destroy the instance on the hypervisor. [ 1097.686467] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1097.687097] env[65121]: DEBUG nova.compute.manager [-] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1097.687202] env[65121]: DEBUG nova.network.neutron [-] [instance: 03a13029-55b1-4757-80a6-9293471434dc] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1097.687651] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.688248] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.688703] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.718109] env[65121]: DEBUG nova.network.neutron [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updated VIF entry in instance network info cache for port 04d47dd2-7f48-48e4-aed4-9d693168b8ca. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1097.718398] env[65121]: DEBUG nova.network.neutron [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": null, "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.780102] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.795774] env[65121]: DEBUG oslo_vmware.api [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107224, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179683} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.796110] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.796298] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1097.796766] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1097.796830] env[65121]: INFO nova.compute.manager [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1097.797203] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1097.797381] env[65121]: DEBUG nova.compute.manager [-] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1097.797489] env[65121]: DEBUG nova.network.neutron [-] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1097.797768] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.798393] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.798688] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.831621] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107220, 'name': ReconfigVM_Task, 'duration_secs': 0.586344} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.831621] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7/3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.832149] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3badf25-89c8-4d10-8354-81183f342427 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.840605] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1097.840605] env[65121]: value = "task-5107226" [ 1097.840605] env[65121]: _type = "Task" [ 1097.840605] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.853989] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107226, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.882497] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107225, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.906169] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.959376] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5225b99d-613f-b7fc-3a24-77845724098f, 'name': SearchDatastore_Task, 'duration_secs': 0.014194} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.959679] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1097.959939] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.960203] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.960353] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.960502] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.960779] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81026f27-5949-401c-b834-bb8b10741c17 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.970856] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.971098] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1097.971923] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f050c105-2a3c-4bc1-9d94-e06fd7dee245 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.979421] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1097.979421] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524a6d60-349c-6b26-92b9-afe62457c284" [ 1097.979421] env[65121]: _type = "Task" [ 1097.979421] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.988980] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524a6d60-349c-6b26-92b9-afe62457c284, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.084809] env[65121]: DEBUG nova.compute.manager [req-403f85ed-4e6e-4b07-98ed-aafac23905d4 req-5b7177ac-1b9b-4a1f-9d1c-7ad036d81a8a service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Received event network-vif-deleted-82e421ec-a3e6-4fcb-a1f4-e88c302e1df2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1098.085045] env[65121]: INFO nova.compute.manager [req-403f85ed-4e6e-4b07-98ed-aafac23905d4 req-5b7177ac-1b9b-4a1f-9d1c-7ad036d81a8a service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Neutron deleted interface 82e421ec-a3e6-4fcb-a1f4-e88c302e1df2; detaching it from the instance and deleting it from the info cache [ 1098.085184] env[65121]: DEBUG nova.network.neutron [req-403f85ed-4e6e-4b07-98ed-aafac23905d4 req-5b7177ac-1b9b-4a1f-9d1c-7ad036d81a8a service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1098.160690] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.166340] env[65121]: DEBUG nova.scheduler.client.report [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1098.222457] env[65121]: DEBUG oslo_concurrency.lockutils [req-b77e31c2-1eff-477d-9c8f-00e2595bdaa8 req-3b4f1ded-53a7-4b56-83c2-7bdb05c19c37 service nova] Releasing lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.271146] env[65121]: DEBUG oslo_concurrency.lockutils [None req-64ee2ce6-91e7-402e-8fd0-2e3a36537723 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.396s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.272382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.112s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.272584] env[65121]: DEBUG nova.compute.manager [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1098.273708] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e09109-5305-4f3c-8c7f-c847ccd5ab2a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.284555] env[65121]: DEBUG nova.compute.manager [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1098.285190] env[65121]: DEBUG nova.objects.instance [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.353886] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107226, 'name': Rename_Task, 'duration_secs': 0.173591} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.354217] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1098.354558] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dec009ce-254b-4ea8-9e0c-f2c19589117c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.362786] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1098.362786] env[65121]: value = "task-5107228" [ 1098.362786] env[65121]: _type = "Task" [ 1098.362786] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.372833] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.385489] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107225, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542602} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.385802] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 9f9118f1-c708-469f-839e-b3c4ef52f38f/9f9118f1-c708-469f-839e-b3c4ef52f38f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1098.386023] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1098.386291] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2ccdd7d-b5a7-4e21-9a6b-5a6c8d30dfca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.394802] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1098.394802] env[65121]: value = "task-5107229" [ 1098.394802] env[65121]: _type = "Task" [ 1098.394802] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.406404] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107229, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.462346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "0497d253-887b-47a9-be8d-cc4592e046ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.462346] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "0497d253-887b-47a9-be8d-cc4592e046ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.493782] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524a6d60-349c-6b26-92b9-afe62457c284, 'name': SearchDatastore_Task, 'duration_secs': 0.015557} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.495164] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0361bd23-3b9f-4e31-8d2e-72fb864b59f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.502845] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1098.502845] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d566e6-adb0-083c-a9ba-53cfa9f848e3" [ 1098.502845] env[65121]: _type = "Task" [ 1098.502845] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.510780] env[65121]: DEBUG nova.compute.manager [req-69bdc29b-680d-48c1-9308-cf0a191993fb req-9a684cd8-b61a-408a-a4d9-3d171170ac73 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Received event network-vif-deleted-4ea2aa90-9370-491b-97b5-2c112a535219 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1098.510864] env[65121]: INFO nova.compute.manager [req-69bdc29b-680d-48c1-9308-cf0a191993fb req-9a684cd8-b61a-408a-a4d9-3d171170ac73 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Neutron deleted interface 4ea2aa90-9370-491b-97b5-2c112a535219; detaching it from the instance and deleting it from the info cache [ 1098.511080] env[65121]: DEBUG nova.network.neutron [req-69bdc29b-680d-48c1-9308-cf0a191993fb req-9a684cd8-b61a-408a-a4d9-3d171170ac73 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1098.516203] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d566e6-adb0-083c-a9ba-53cfa9f848e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.521681] env[65121]: DEBUG nova.network.neutron [-] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1098.588141] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b7c3975-4b91-4380-bd50-3c59fef3c499 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.601342] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ed16b4-c042-4f60-ba6f-e28639d32e69 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.646288] env[65121]: DEBUG nova.compute.manager [req-403f85ed-4e6e-4b07-98ed-aafac23905d4 req-5b7177ac-1b9b-4a1f-9d1c-7ad036d81a8a service nova] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Detach interface failed, port_id=82e421ec-a3e6-4fcb-a1f4-e88c302e1df2, reason: Instance 03a13029-55b1-4757-80a6-9293471434dc could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1098.671823] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.674313] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.068s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.674567] env[65121]: DEBUG nova.objects.instance [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lazy-loading 'resources' on Instance uuid 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.696102] env[65121]: INFO nova.scheduler.client.report [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Deleted allocations for instance 347bbb23-da89-4e6f-837c-e88bf7c57a3a [ 1098.874790] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107228, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.904271] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107229, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072044} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.904516] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1098.905329] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fa135c-0e93-4d8c-b562-82c989e88ae8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.928353] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 9f9118f1-c708-469f-839e-b3c4ef52f38f/9f9118f1-c708-469f-839e-b3c4ef52f38f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.928617] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbfc03ed-53d9-4bdb-807f-6ddfe172dc8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.949846] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1098.949846] env[65121]: value = "task-5107230" [ 1098.949846] env[65121]: _type = "Task" [ 1098.949846] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.959608] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107230, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.965186] env[65121]: DEBUG nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1098.987241] env[65121]: DEBUG nova.network.neutron [-] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1099.015103] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d566e6-adb0-083c-a9ba-53cfa9f848e3, 'name': SearchDatastore_Task, 'duration_secs': 0.023034} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.016412] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.016412] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1099.016412] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ebc6145-0370-4799-93b8-6a077b093c3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.019702] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7dce4aa-bb6c-4cea-926b-19efb02f1200 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.024708] env[65121]: INFO nova.compute.manager [-] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Took 1.34 seconds to deallocate network for instance. [ 1099.029816] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1099.029816] env[65121]: value = "task-5107231" [ 1099.029816] env[65121]: _type = "Task" [ 1099.029816] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.035592] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141e74ec-9c52-44d6-b015-699cf0595da9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.057522] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.078872] env[65121]: DEBUG nova.compute.manager [req-69bdc29b-680d-48c1-9308-cf0a191993fb req-9a684cd8-b61a-408a-a4d9-3d171170ac73 service nova] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Detach interface failed, port_id=4ea2aa90-9370-491b-97b5-2c112a535219, reason: Instance c4096314-270a-4270-9e1d-5ace8ddbd286 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1099.178029] env[65121]: DEBUG nova.objects.instance [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lazy-loading 'numa_topology' on Instance uuid 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.207944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0d1672af-427b-4f86-bff3-660d7ff027f2 tempest-AttachInterfacesV270Test-1517508842 tempest-AttachInterfacesV270Test-1517508842-project-member] Lock "347bbb23-da89-4e6f-837c-e88bf7c57a3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.752s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.294556] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.294932] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7484a1d9-202a-451a-abc4-f43bfe89c0ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.304143] env[65121]: DEBUG oslo_vmware.api [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1099.304143] env[65121]: value = "task-5107233" [ 1099.304143] env[65121]: _type = "Task" [ 1099.304143] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.315371] env[65121]: DEBUG oslo_vmware.api [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107233, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.377145] env[65121]: DEBUG oslo_vmware.api [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107228, 'name': PowerOnVM_Task, 'duration_secs': 0.858701} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.377313] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1099.377532] env[65121]: INFO nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Took 9.56 seconds to spawn the instance on the hypervisor. [ 1099.377690] env[65121]: DEBUG nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1099.378674] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cad4935-f588-483e-be04-ce71c2dc0e6e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.467530] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107230, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.491041] env[65121]: INFO nova.compute.manager [-] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Took 1.69 seconds to deallocate network for instance. [ 1099.497622] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1099.549852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1099.561822] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107231, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.683428] env[65121]: DEBUG nova.objects.base [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Object Instance<52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731> lazy-loaded attributes: resources,numa_topology {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1099.814585] env[65121]: DEBUG oslo_vmware.api [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107233, 'name': PowerOffVM_Task, 'duration_secs': 0.272651} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.817271] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.817505] env[65121]: DEBUG nova.compute.manager [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1099.818486] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196c4fe3-1f47-4ce3-99e3-bfbe30ee97c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.904874] env[65121]: INFO nova.compute.manager [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Took 23.27 seconds to build instance. [ 1099.912145] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cab324d-6ef2-4a58-8592-668e7b52af93 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.924673] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf1e7c3-7b57-46a0-96a2-b23369acf68d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.960343] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa58e5b-2b2c-4ebb-8043-30411f067310 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.973442] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87da013d-66d9-4384-bee7-4a5ef54356b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.977343] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107230, 'name': ReconfigVM_Task, 'duration_secs': 0.52591} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.977979] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 9f9118f1-c708-469f-839e-b3c4ef52f38f/9f9118f1-c708-469f-839e-b3c4ef52f38f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.979045] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c83311d8-353e-4ed0-9033-86a289111e50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.989719] env[65121]: DEBUG nova.compute.provider_tree [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.992531] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1099.992531] env[65121]: value = "task-5107234" [ 1099.992531] env[65121]: _type = "Task" [ 1099.992531] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.004126] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.005607] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107234, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.059427] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107231, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693314} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.060116] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1100.060116] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1100.060334] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8fcc3995-5411-4811-a097-2583e0fd668e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.068125] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1100.068125] env[65121]: value = "task-5107235" [ 1100.068125] env[65121]: _type = "Task" [ 1100.068125] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.079174] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107235, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.332549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ce1c2f2f-1bd9-49d5-836b-151d87344398 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.060s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.360143] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1100.360425] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1100.361389] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d17a07-da16-4014-b4c4-57a810f187c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.378855] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c755db70-3069-46d6-9600-b9022f1e9a2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.407054] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] volume-41abb581-3841-4c0e-b79e-9820e1b241f1/volume-41abb581-3841-4c0e-b79e-9820e1b241f1.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1100.407618] env[65121]: DEBUG oslo_concurrency.lockutils [None req-32644bf1-df42-4e5b-af93-41715ed8bb40 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.779s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.407914] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88c87b37-c716-4e63-8d4b-da0705e21122 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.430238] env[65121]: DEBUG oslo_vmware.api [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1100.430238] env[65121]: value = "task-5107236" [ 1100.430238] env[65121]: _type = "Task" [ 1100.430238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.439418] env[65121]: DEBUG oslo_vmware.api [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107236, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.494468] env[65121]: DEBUG nova.scheduler.client.report [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1100.513513] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107234, 'name': Rename_Task, 'duration_secs': 0.378592} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.514987] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.515448] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da17c7d9-7550-4464-b91a-5806877cb51a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.525667] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1100.525667] env[65121]: value = "task-5107237" [ 1100.525667] env[65121]: _type = "Task" [ 1100.525667] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.541702] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.583929] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107235, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107869} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.583929] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1100.584297] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e9d354-d00e-451b-990a-4e186a57a2f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.608958] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1100.609736] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11445513-3dbe-4ece-9492-b2c490b505d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.634550] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1100.634550] env[65121]: value = "task-5107238" [ 1100.634550] env[65121]: _type = "Task" [ 1100.634550] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.648431] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107238, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.942794] env[65121]: DEBUG oslo_vmware.api [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107236, 'name': ReconfigVM_Task, 'duration_secs': 0.429716} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.943112] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfigured VM instance instance-0000005c to attach disk [datastore2] volume-41abb581-3841-4c0e-b79e-9820e1b241f1/volume-41abb581-3841-4c0e-b79e-9820e1b241f1.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.951468] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03321335-e728-449c-b347-19d0d2c4cc27 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.969222] env[65121]: DEBUG oslo_vmware.api [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1100.969222] env[65121]: value = "task-5107239" [ 1100.969222] env[65121]: _type = "Task" [ 1100.969222] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.978892] env[65121]: DEBUG oslo_vmware.api [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.006045] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.332s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.009083] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.359s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.010757] env[65121]: INFO nova.compute.claims [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1101.041230] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107237, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.142768] env[65121]: DEBUG nova.compute.manager [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1101.142963] env[65121]: DEBUG nova.compute.manager [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing instance network info cache due to event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1101.143193] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.143377] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.143478] env[65121]: DEBUG nova.network.neutron [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1101.149663] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107238, 'name': ReconfigVM_Task, 'duration_secs': 0.342093} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.150261] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 39ca7a89-c262-4169-91b1-92ea45744518/39ca7a89-c262-4169-91b1-92ea45744518.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1101.150885] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-770fb75e-6c5f-4c55-83fb-ffcd23684a38 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.159037] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1101.159037] env[65121]: value = "task-5107240" [ 1101.159037] env[65121]: _type = "Task" [ 1101.159037] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.168153] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107240, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.482050] env[65121]: DEBUG oslo_vmware.api [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107239, 'name': ReconfigVM_Task, 'duration_secs': 0.18802} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.482537] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1101.522186] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f3d19690-0a19-4a63-b880-13e24e72c391 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.912s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.523106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.863s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.523294] env[65121]: INFO nova.compute.manager [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Unshelving [ 1101.543764] env[65121]: DEBUG oslo_vmware.api [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107237, 'name': PowerOnVM_Task, 'duration_secs': 0.58785} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.543909] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.544128] env[65121]: INFO nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Took 9.23 seconds to spawn the instance on the hypervisor. [ 1101.544304] env[65121]: DEBUG nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1101.545605] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e197a8-c69f-4ab4-a89c-f856d6319db0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.650807] env[65121]: WARNING neutronclient.v2_0.client [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.651636] env[65121]: WARNING openstack [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.652163] env[65121]: WARNING openstack [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.670979] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107240, 'name': Rename_Task, 'duration_secs': 0.215973} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.671279] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1101.671678] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cccf1ba8-9066-4801-a5d8-7182ca91aefa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.680418] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1101.680418] env[65121]: value = "task-5107242" [ 1101.680418] env[65121]: _type = "Task" [ 1101.680418] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.687821] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.745847] env[65121]: DEBUG nova.objects.instance [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.752533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.752765] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.813355] env[65121]: WARNING openstack [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.813851] env[65121]: WARNING openstack [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.900017] env[65121]: WARNING neutronclient.v2_0.client [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.900783] env[65121]: WARNING openstack [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.901366] env[65121]: WARNING openstack [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.005014] env[65121]: DEBUG nova.network.neutron [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updated VIF entry in instance network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1102.005349] env[65121]: DEBUG nova.network.neutron [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1102.066445] env[65121]: INFO nova.compute.manager [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Took 23.81 seconds to build instance. [ 1102.189893] env[65121]: DEBUG oslo_vmware.api [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107242, 'name': PowerOnVM_Task, 'duration_secs': 0.506703} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.190178] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1102.190990] env[65121]: DEBUG nova.compute.manager [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1102.191274] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532b5648-24ed-495d-9548-aedcad59152f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.254949] env[65121]: DEBUG nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1102.262983] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.263178] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.263344] env[65121]: DEBUG nova.network.neutron [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1102.263507] env[65121]: DEBUG nova.objects.instance [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'info_cache' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.334541] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c169594-f460-43bf-a1f4-7810d7ca70ea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.344298] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cae8b5-84a2-47d8-a75a-93bb3c0545e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.380749] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53c662d-d7d4-4d3c-b87f-eef41a9bed8d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.390682] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab091cd-66aa-403c-afa3-f6cb4683269e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.405814] env[65121]: DEBUG nova.compute.provider_tree [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.507911] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f6d6c97-9505-4b2f-8e89-082ecb49e63d req-5ce269a0-a4f1-48ef-9e0d-6356d9e4dc32 service nova] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.539358] env[65121]: DEBUG nova.objects.instance [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'flavor' on Instance uuid 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.559338] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.569242] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520ca08a-6be2-4a20-adba-a5375bffd34b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.324s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.711491] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.769415] env[65121]: DEBUG nova.objects.base [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Object Instance<6fe86deb-0903-4769-a05c-b7d0acec0103> lazy-loaded attributes: flavor,info_cache {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1102.785499] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.909325] env[65121]: DEBUG nova.scheduler.client.report [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1102.930228] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "39ca7a89-c262-4169-91b1-92ea45744518" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.930498] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "39ca7a89-c262-4169-91b1-92ea45744518" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.930698] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "39ca7a89-c262-4169-91b1-92ea45744518-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.930907] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "39ca7a89-c262-4169-91b1-92ea45744518-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.931088] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "39ca7a89-c262-4169-91b1-92ea45744518-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.935205] env[65121]: INFO nova.compute.manager [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Terminating instance [ 1103.050432] env[65121]: DEBUG oslo_concurrency.lockutils [None req-05a176ad-fa5c-4cbf-aa40-e4caa37ee89e tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.338s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1103.279252] env[65121]: WARNING neutronclient.v2_0.client [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.279252] env[65121]: WARNING openstack [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.279252] env[65121]: WARNING openstack [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.416726] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1103.417639] env[65121]: DEBUG nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1103.420921] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.924s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1103.424491] env[65121]: INFO nova.compute.claims [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1103.442051] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "refresh_cache-39ca7a89-c262-4169-91b1-92ea45744518" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.442223] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquired lock "refresh_cache-39ca7a89-c262-4169-91b1-92ea45744518" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.442404] env[65121]: DEBUG nova.network.neutron [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1103.530811] env[65121]: WARNING openstack [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.531214] env[65121]: WARNING openstack [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.626167] env[65121]: WARNING neutronclient.v2_0.client [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.626923] env[65121]: WARNING openstack [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.627292] env[65121]: WARNING openstack [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.698838] env[65121]: INFO nova.compute.manager [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Rebuilding instance [ 1103.735146] env[65121]: DEBUG nova.network.neutron [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating instance_info_cache with network_info: [{"id": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "address": "fa:16:3e:0e:bf:f3", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape31cbeae-05", "ovs_interfaceid": "e31cbeae-05bc-416d-9df3-10f09d947ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1103.765146] env[65121]: DEBUG nova.compute.manager [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1103.767372] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62b2a63-0689-41dc-895d-efa7caeb6dce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.928078] env[65121]: DEBUG nova.compute.utils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1103.932286] env[65121]: DEBUG nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1103.932380] env[65121]: DEBUG nova.network.neutron [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1103.932660] env[65121]: WARNING neutronclient.v2_0.client [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.935070] env[65121]: WARNING neutronclient.v2_0.client [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.935070] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.935070] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.956170] env[65121]: WARNING neutronclient.v2_0.client [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.958207] env[65121]: WARNING openstack [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.958207] env[65121]: WARNING openstack [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.000813] env[65121]: DEBUG nova.policy [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26beb81d28f44cc8932ea8e87a4cd2c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0630960dcbf44781be05184565d81932', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1104.013216] env[65121]: DEBUG nova.network.neutron [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1104.121639] env[65121]: DEBUG nova.network.neutron [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.243271] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "refresh_cache-6fe86deb-0903-4769-a05c-b7d0acec0103" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.278575] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64fee846-6885-48bb-b3d4-2c740716afc4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.293712] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff9bfa6-f72a-4c74-acf2-646880f5ac98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.329925] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8cfa20-620d-4f13-bf20-915cd4d3065b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.339920] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54597c1-f097-4a26-93ec-78a989692fcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.356972] env[65121]: DEBUG nova.compute.provider_tree [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.390723] env[65121]: DEBUG nova.network.neutron [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Successfully created port: 98d8432f-dc60-460d-96b2-2c83d89b9c78 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1104.407812] env[65121]: DEBUG nova.compute.manager [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1104.431388] env[65121]: DEBUG nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1104.628086] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Releasing lock "refresh_cache-39ca7a89-c262-4169-91b1-92ea45744518" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.628583] env[65121]: DEBUG nova.compute.manager [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1104.629367] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1104.630319] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1bc893-c41a-49ea-9b9e-8eedd82a47d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.639217] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1104.639508] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-903ac04a-8db9-4437-950f-caa07f6529fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.647505] env[65121]: DEBUG oslo_vmware.api [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1104.647505] env[65121]: value = "task-5107244" [ 1104.647505] env[65121]: _type = "Task" [ 1104.647505] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.657048] env[65121]: DEBUG oslo_vmware.api [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.788275] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1104.788906] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e092634-8d6e-4a35-9bbc-77568085d1ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.799408] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1104.799408] env[65121]: value = "task-5107245" [ 1104.799408] env[65121]: _type = "Task" [ 1104.799408] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.809335] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.860931] env[65121]: DEBUG nova.scheduler.client.report [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1104.929880] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.126106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "714930d8-3ec9-4aa8-831a-a9f28c365545" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.126426] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.158768] env[65121]: DEBUG oslo_vmware.api [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107244, 'name': PowerOffVM_Task, 'duration_secs': 0.204035} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.159088] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1105.159325] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1105.159628] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1cd7ddf8-f0f5-4af4-9db1-b89d79a05b0f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.185695] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1105.185991] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1105.186804] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Deleting the datastore file [datastore2] 39ca7a89-c262-4169-91b1-92ea45744518 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.186804] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5091c375-e63a-4d48-a53f-3228629d3692 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.193819] env[65121]: DEBUG oslo_vmware.api [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for the task: (returnval){ [ 1105.193819] env[65121]: value = "task-5107247" [ 1105.193819] env[65121]: _type = "Task" [ 1105.193819] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.203646] env[65121]: DEBUG oslo_vmware.api [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.251167] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1105.251550] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d080d9ae-fdc8-4b12-bd3d-db3fc934ef3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.259429] env[65121]: DEBUG oslo_vmware.api [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1105.259429] env[65121]: value = "task-5107248" [ 1105.259429] env[65121]: _type = "Task" [ 1105.259429] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.267844] env[65121]: DEBUG oslo_vmware.api [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107248, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.312415] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107245, 'name': PowerOffVM_Task, 'duration_secs': 0.202739} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.312777] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1105.368014] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.947s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.369019] env[65121]: DEBUG nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1105.371776] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.822s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.372698] env[65121]: DEBUG nova.objects.instance [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lazy-loading 'resources' on Instance uuid 03a13029-55b1-4757-80a6-9293471434dc {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.395262] env[65121]: INFO nova.compute.manager [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Detaching volume 41abb581-3841-4c0e-b79e-9820e1b241f1 [ 1105.435301] env[65121]: INFO nova.virt.block_device [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Attempting to driver detach volume 41abb581-3841-4c0e-b79e-9820e1b241f1 from mountpoint /dev/sdb [ 1105.435656] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1105.435868] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1105.436823] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16c7715-5734-4c8b-ad5b-6df5925c1ce8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.440714] env[65121]: DEBUG nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1105.471919] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa8c0fc-5d01-4bcc-9ad0-6f04a88e0242 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.482668] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1105.484945] env[65121]: DEBUG nova.virt.hardware [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1105.485745] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951cabf1-9a48-469a-88a2-300d8e8949ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.488996] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcd3b19-0d7d-489c-887d-e4f8eb05db06 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.517773] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917a7fd5-0040-4397-9a00-6717370a0abf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.522833] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e924d04d-fd05-4e7f-a9b5-ea5069abb652 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.552694] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The volume has not been displaced from its original location: [datastore2] volume-41abb581-3841-4c0e-b79e-9820e1b241f1/volume-41abb581-3841-4c0e-b79e-9820e1b241f1.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1105.556899] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1105.557489] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7337998-f6eb-4869-8ffe-1b842f2bef22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.577200] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1105.577200] env[65121]: value = "task-5107250" [ 1105.577200] env[65121]: _type = "Task" [ 1105.577200] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.588252] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107250, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.629224] env[65121]: DEBUG nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1105.706312] env[65121]: DEBUG oslo_vmware.api [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Task: {'id': task-5107247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184427} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.706312] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.706635] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1105.706635] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.706742] env[65121]: INFO nova.compute.manager [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1105.707976] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1105.707976] env[65121]: DEBUG nova.compute.manager [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1105.707976] env[65121]: DEBUG nova.network.neutron [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1105.707976] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.708506] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.708856] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.765362] env[65121]: DEBUG nova.network.neutron [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1105.765892] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.776372] env[65121]: DEBUG oslo_vmware.api [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107248, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.876260] env[65121]: DEBUG nova.compute.utils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1105.883971] env[65121]: DEBUG nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1105.884942] env[65121]: DEBUG nova.network.neutron [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1105.885525] env[65121]: WARNING neutronclient.v2_0.client [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.886510] env[65121]: WARNING neutronclient.v2_0.client [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.890140] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1105.890140] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1105.954044] env[65121]: DEBUG nova.compute.manager [req-37d327db-cde5-492e-85c4-1fdc330eba39 req-aef3ec09-fe6a-43db-9f8e-e4c05632c935 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Received event network-vif-plugged-98d8432f-dc60-460d-96b2-2c83d89b9c78 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1105.954439] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d327db-cde5-492e-85c4-1fdc330eba39 req-aef3ec09-fe6a-43db-9f8e-e4c05632c935 service nova] Acquiring lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.954826] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d327db-cde5-492e-85c4-1fdc330eba39 req-aef3ec09-fe6a-43db-9f8e-e4c05632c935 service nova] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.955167] env[65121]: DEBUG oslo_concurrency.lockutils [req-37d327db-cde5-492e-85c4-1fdc330eba39 req-aef3ec09-fe6a-43db-9f8e-e4c05632c935 service nova] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.955452] env[65121]: DEBUG nova.compute.manager [req-37d327db-cde5-492e-85c4-1fdc330eba39 req-aef3ec09-fe6a-43db-9f8e-e4c05632c935 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] No waiting events found dispatching network-vif-plugged-98d8432f-dc60-460d-96b2-2c83d89b9c78 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1105.958035] env[65121]: WARNING nova.compute.manager [req-37d327db-cde5-492e-85c4-1fdc330eba39 req-aef3ec09-fe6a-43db-9f8e-e4c05632c935 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Received unexpected event network-vif-plugged-98d8432f-dc60-460d-96b2-2c83d89b9c78 for instance with vm_state building and task_state spawning. [ 1105.959898] env[65121]: DEBUG nova.policy [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1c51fc43b7f44ce9b023d8cb769e029', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89110618b6ef4ce68b2ca53e7e168139', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1106.094393] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107250, 'name': ReconfigVM_Task, 'duration_secs': 0.302529} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.097939] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1106.104817] env[65121]: DEBUG nova.network.neutron [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Successfully updated port: 98d8432f-dc60-460d-96b2-2c83d89b9c78 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1106.107510] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f62b7dc4-0a26-4298-a510-28baf35865bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.132165] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1106.132165] env[65121]: value = "task-5107251" [ 1106.132165] env[65121]: _type = "Task" [ 1106.132165] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.153639] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107251, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.166191] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.268491] env[65121]: DEBUG nova.network.neutron [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1106.276629] env[65121]: DEBUG oslo_vmware.api [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107248, 'name': PowerOnVM_Task, 'duration_secs': 0.531437} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.276984] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.277061] env[65121]: DEBUG nova.compute.manager [None req-3e6656d6-5f78-4002-ab69-f2b387ee1e62 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1106.278103] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a54a1dc-58cd-4686-ae39-fccb74d46d8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.295790] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74b6c9c-d7b7-464b-ae1f-9c07bc0b2fe8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.305185] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304f4eb4-ea19-40c2-9ad4-2f1353f707d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.342319] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6748fe22-d274-4701-afb9-e0d5ace160da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.353235] env[65121]: DEBUG nova.network.neutron [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Successfully created port: d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1106.359435] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0722e127-3441-4088-9e30-0271a5ca17a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.375376] env[65121]: DEBUG nova.compute.provider_tree [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.398931] env[65121]: DEBUG nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1106.585554] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.585554] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.625305] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.625488] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.625677] env[65121]: DEBUG nova.network.neutron [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1106.643339] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107251, 'name': ReconfigVM_Task, 'duration_secs': 0.285912} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.643667] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1106.771431] env[65121]: INFO nova.compute.manager [-] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Took 1.06 seconds to deallocate network for instance. [ 1106.879049] env[65121]: DEBUG nova.scheduler.client.report [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1107.094670] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.094900] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.095312] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.095312] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.095454] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.095829] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.095829] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1107.095964] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.128724] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.129159] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.167422] env[65121]: DEBUG nova.network.neutron [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1107.202211] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.202624] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.272124] env[65121]: WARNING neutronclient.v2_0.client [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.272817] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.273191] env[65121]: WARNING openstack [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.282342] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1107.384300] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.012s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.387489] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.383s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1107.387489] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.390402] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.831s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1107.390527] env[65121]: DEBUG nova.objects.instance [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lazy-loading 'pci_requests' on Instance uuid 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.410653] env[65121]: DEBUG nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1107.414101] env[65121]: INFO nova.scheduler.client.report [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted allocations for instance c4096314-270a-4270-9e1d-5ace8ddbd286 [ 1107.419865] env[65121]: INFO nova.scheduler.client.report [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Deleted allocations for instance 03a13029-55b1-4757-80a6-9293471434dc [ 1107.448245] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1107.448713] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1107.448713] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1107.449667] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1107.450239] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1107.450239] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1107.450239] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.450702] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1107.450702] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1107.450825] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1107.450961] env[65121]: DEBUG nova.virt.hardware [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1107.452059] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabadac8-b975-4b63-9889-2fefc7a5d28a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.464029] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96551004-bb34-493d-8158-0b570ce3c5a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.521416] env[65121]: DEBUG nova.network.neutron [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Updating instance_info_cache with network_info: [{"id": "98d8432f-dc60-460d-96b2-2c83d89b9c78", "address": "fa:16:3e:6d:2a:fc", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98d8432f-dc", "ovs_interfaceid": "98d8432f-dc60-460d-96b2-2c83d89b9c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1107.600185] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1107.715567] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1107.715996] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa18ad54-7454-48bc-976d-53ba20136f6c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.724786] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1107.724786] env[65121]: value = "task-5107253" [ 1107.724786] env[65121]: _type = "Task" [ 1107.724786] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.736287] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.895120] env[65121]: DEBUG nova.objects.instance [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lazy-loading 'numa_topology' on Instance uuid 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.927106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-13f6ed57-dff5-4b7a-86d7-b7c2a7896b62 tempest-ServerRescueTestJSONUnderV235-479719765 tempest-ServerRescueTestJSONUnderV235-479719765-project-member] Lock "03a13029-55b1-4757-80a6-9293471434dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.953s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.928725] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d6e0cb65-894c-405e-bfe2-8f40e53fe6b0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "c4096314-270a-4270-9e1d-5ace8ddbd286" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.816s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.998103] env[65121]: DEBUG nova.compute.manager [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Received event network-changed-98d8432f-dc60-460d-96b2-2c83d89b9c78 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1107.998263] env[65121]: DEBUG nova.compute.manager [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Refreshing instance network info cache due to event network-changed-98d8432f-dc60-460d-96b2-2c83d89b9c78. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1107.998450] env[65121]: DEBUG oslo_concurrency.lockutils [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Acquiring lock "refresh_cache-e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.000512] env[65121]: DEBUG nova.network.neutron [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Successfully updated port: d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1108.024915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.025395] env[65121]: DEBUG nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Instance network_info: |[{"id": "98d8432f-dc60-460d-96b2-2c83d89b9c78", "address": "fa:16:3e:6d:2a:fc", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98d8432f-dc", "ovs_interfaceid": "98d8432f-dc60-460d-96b2-2c83d89b9c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1108.025945] env[65121]: DEBUG oslo_concurrency.lockutils [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Acquired lock "refresh_cache-e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.026433] env[65121]: DEBUG nova.network.neutron [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Refreshing network info cache for port 98d8432f-dc60-460d-96b2-2c83d89b9c78 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1108.027947] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:2a:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '98d8432f-dc60-460d-96b2-2c83d89b9c78', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.039450] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1108.041124] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1108.041810] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b03f60fc-1219-4c7a-af5d-85b5787efad0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.070301] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1108.070301] env[65121]: value = "task-5107254" [ 1108.070301] env[65121]: _type = "Task" [ 1108.070301] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.082391] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107254, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.238046] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1108.238289] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1108.238480] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1108.239514] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6ded29-f653-4b7b-b848-5793db2748e9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.260743] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab934541-4be7-46b6-9008-2ef210e75236 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.268291] env[65121]: WARNING nova.virt.vmwareapi.driver [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1108.268636] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1108.269799] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1acca35a-b657-4350-b30a-ce86e700ad2a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.277928] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1108.278216] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7423d319-7c2d-464a-a81a-d955e65b6389 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.366081] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.366655] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.368919] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1108.369259] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1108.369549] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleting the datastore file [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.370315] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfe94d98-f62f-4c84-867b-61d5d7fbc6a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.380963] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1108.380963] env[65121]: value = "task-5107256" [ 1108.380963] env[65121]: _type = "Task" [ 1108.380963] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.392222] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107256, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.401710] env[65121]: INFO nova.compute.claims [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1108.504064] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "refresh_cache-0497d253-887b-47a9-be8d-cc4592e046ba" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.504855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "refresh_cache-0497d253-887b-47a9-be8d-cc4592e046ba" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.504855] env[65121]: DEBUG nova.network.neutron [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1108.540736] env[65121]: WARNING neutronclient.v2_0.client [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.541477] env[65121]: WARNING openstack [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.541895] env[65121]: WARNING openstack [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.582353] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107254, 'name': CreateVM_Task, 'duration_secs': 0.46006} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.582596] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1108.583223] env[65121]: WARNING neutronclient.v2_0.client [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.583813] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.583985] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1108.584358] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1108.585068] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b36dea89-a37a-40ec-a356-f278c2206001 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.591656] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1108.591656] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528ad786-5951-9390-b744-617f204f1cc0" [ 1108.591656] env[65121]: _type = "Task" [ 1108.591656] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.601729] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528ad786-5951-9390-b744-617f204f1cc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.772726] env[65121]: WARNING openstack [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.773129] env[65121]: WARNING openstack [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.872221] env[65121]: INFO nova.compute.manager [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Detaching volume 02de1e93-4004-47b3-a1b4-d2690ef0da94 [ 1108.877307] env[65121]: WARNING neutronclient.v2_0.client [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.878393] env[65121]: WARNING openstack [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.878960] env[65121]: WARNING openstack [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.903057] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107256, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.380151} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.903892] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1108.903892] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1108.903892] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1108.928522] env[65121]: INFO nova.virt.block_device [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Attempting to driver detach volume 02de1e93-4004-47b3-a1b4-d2690ef0da94 from mountpoint /dev/sdb [ 1108.928679] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1108.928893] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993500', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'name': 'volume-02de1e93-4004-47b3-a1b4-d2690ef0da94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'e0d3e3c8-6471-4345-8677-369612674769', 'attached_at': '', 'detached_at': '', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'serial': '02de1e93-4004-47b3-a1b4-d2690ef0da94'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1108.929842] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a92c155-d6fe-48e9-b0e1-d4b41609bf6f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.961707] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29622f4-17a0-4fd5-895a-3b27fb65caee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.971548] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47a24e9-3eb3-42fa-a9d4-693271bf5b25 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.996346] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304f28b4-07f0-4d0f-987b-1235fe0ced0f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.013920] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.014425] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.023142] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The volume has not been displaced from its original location: [datastore1] volume-02de1e93-4004-47b3-a1b4-d2690ef0da94/volume-02de1e93-4004-47b3-a1b4-d2690ef0da94.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1109.028927] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfiguring VM instance instance-0000003a to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1109.029569] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81ba06c9-a9d5-4a2d-b0ad-e69518a51f24 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.058667] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1109.058667] env[65121]: value = "task-5107257" [ 1109.058667] env[65121]: _type = "Task" [ 1109.058667] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.067963] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107257, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.076946] env[65121]: DEBUG nova.network.neutron [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Updated VIF entry in instance network info cache for port 98d8432f-dc60-460d-96b2-2c83d89b9c78. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1109.077837] env[65121]: DEBUG nova.network.neutron [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Updating instance_info_cache with network_info: [{"id": "98d8432f-dc60-460d-96b2-2c83d89b9c78", "address": "fa:16:3e:6d:2a:fc", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98d8432f-dc", "ovs_interfaceid": "98d8432f-dc60-460d-96b2-2c83d89b9c78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.104881] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528ad786-5951-9390-b744-617f204f1cc0, 'name': SearchDatastore_Task, 'duration_secs': 0.011646} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.105636] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.105636] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1109.105801] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.105972] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.106267] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1109.106441] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-552423e3-9b2b-4787-803b-2c36defd51c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.116529] env[65121]: DEBUG nova.network.neutron [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1109.122573] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1109.122733] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1109.123469] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1246e8e-d338-4b77-9553-e3ad6666e1df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.130276] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1109.130276] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fb07a1-3e28-82e5-76ea-2b68991f6ca7" [ 1109.130276] env[65121]: _type = "Task" [ 1109.130276] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.143375] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fb07a1-3e28-82e5-76ea-2b68991f6ca7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.190327] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.190721] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.361802] env[65121]: WARNING neutronclient.v2_0.client [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.362613] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.363684] env[65121]: WARNING openstack [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.410082] env[65121]: INFO nova.virt.block_device [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Booting with volume 41abb581-3841-4c0e-b79e-9820e1b241f1 at /dev/sdb [ 1109.481395] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdfebf10-7d9c-4248-91db-8515cdabcb90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.491144] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1449d9eb-3bb0-46fa-8447-937bdf14190a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.534885] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f2ce57f-223a-479b-90e0-84ebca8c3930 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.546299] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf508f3-d57a-4e8b-b8c7-9c80876ee4ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.561302] env[65121]: DEBUG nova.network.neutron [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Updating instance_info_cache with network_info: [{"id": "d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d", "address": "fa:16:3e:04:e9:c3", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3da86fe-ad", "ovs_interfaceid": "d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.571665] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107257, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.588362] env[65121]: DEBUG oslo_concurrency.lockutils [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Releasing lock "refresh_cache-e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.588511] env[65121]: DEBUG nova.compute.manager [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Received event network-vif-plugged-d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1109.588666] env[65121]: DEBUG oslo_concurrency.lockutils [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Acquiring lock "0497d253-887b-47a9-be8d-cc4592e046ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.589056] env[65121]: DEBUG oslo_concurrency.lockutils [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Lock "0497d253-887b-47a9-be8d-cc4592e046ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.589056] env[65121]: DEBUG oslo_concurrency.lockutils [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] Lock "0497d253-887b-47a9-be8d-cc4592e046ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.589175] env[65121]: DEBUG nova.compute.manager [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] No waiting events found dispatching network-vif-plugged-d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1109.589524] env[65121]: WARNING nova.compute.manager [req-01e2be77-7b74-4800-9739-a5d1bf24fc74 req-46b97fb2-140a-4edb-a5bc-dec414374a24 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Received unexpected event network-vif-plugged-d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d for instance with vm_state building and task_state spawning. [ 1109.590974] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149a4a84-9efa-4bd1-8840-cf2d10d4b080 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.602148] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5de6e4-8be7-472f-974e-bb848edfee44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.617557] env[65121]: DEBUG nova.virt.block_device [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updating existing volume attachment record: 166c385e-823d-4e12-bd55-d0b6f8b4633e {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1109.642480] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fb07a1-3e28-82e5-76ea-2b68991f6ca7, 'name': SearchDatastore_Task, 'duration_secs': 0.014228} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.645895] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48860fbd-5b2f-42a4-96bf-a8bf947cc7af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.651697] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1109.651697] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e98b09-571f-706e-0d63-92b4ba2906a7" [ 1109.651697] env[65121]: _type = "Task" [ 1109.651697] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.663791] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e98b09-571f-706e-0d63-92b4ba2906a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.770180] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a004427-8ec3-4763-a55e-e518abc3a42e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.779609] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5a5294-bf20-4d53-b393-db378b3b8d86 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.814257] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c07edd-971e-4f4a-8e87-c7083c09312c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.823938] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99a7530-6cab-497b-92e2-1d46daf23ae9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.840997] env[65121]: DEBUG nova.compute.provider_tree [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.031014] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1110.031261] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1110.033645] env[65121]: DEBUG nova.compute.manager [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Received event network-changed-d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1110.033908] env[65121]: DEBUG nova.compute.manager [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Refreshing instance network info cache due to event network-changed-d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1110.034152] env[65121]: DEBUG oslo_concurrency.lockutils [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Acquiring lock "refresh_cache-0497d253-887b-47a9-be8d-cc4592e046ba" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.063576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "refresh_cache-0497d253-887b-47a9-be8d-cc4592e046ba" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.063944] env[65121]: DEBUG nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Instance network_info: |[{"id": "d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d", "address": "fa:16:3e:04:e9:c3", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3da86fe-ad", "ovs_interfaceid": "d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1110.067576] env[65121]: DEBUG oslo_concurrency.lockutils [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Acquired lock "refresh_cache-0497d253-887b-47a9-be8d-cc4592e046ba" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1110.067926] env[65121]: DEBUG nova.network.neutron [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Refreshing network info cache for port d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1110.069577] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:e9:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.077256] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1110.079291] env[65121]: WARNING neutronclient.v2_0.client [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.079916] env[65121]: WARNING openstack [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.080514] env[65121]: WARNING openstack [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.087431] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1110.088546] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-937b65f2-5204-4dd0-a66a-da6498674cb8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.107474] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107257, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.113797] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.113797] env[65121]: value = "task-5107258" [ 1110.113797] env[65121]: _type = "Task" [ 1110.113797] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.123507] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107258, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.163454] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e98b09-571f-706e-0d63-92b4ba2906a7, 'name': SearchDatastore_Task, 'duration_secs': 0.011027} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.163752] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.164350] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2/e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1110.164350] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b02c654-31e7-4d9e-9c20-1131c22d6c1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.171790] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1110.171790] env[65121]: value = "task-5107259" [ 1110.171790] env[65121]: _type = "Task" [ 1110.171790] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.184078] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.209046] env[65121]: WARNING openstack [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.209428] env[65121]: WARNING openstack [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.289402] env[65121]: WARNING neutronclient.v2_0.client [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.290431] env[65121]: WARNING openstack [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.292009] env[65121]: WARNING openstack [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.345023] env[65121]: DEBUG nova.scheduler.client.report [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1110.461068] env[65121]: DEBUG nova.network.neutron [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Updated VIF entry in instance network info cache for port d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1110.461592] env[65121]: DEBUG nova.network.neutron [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Updating instance_info_cache with network_info: [{"id": "d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d", "address": "fa:16:3e:04:e9:c3", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3da86fe-ad", "ovs_interfaceid": "d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1110.535433] env[65121]: DEBUG nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1110.576928] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107257, 'name': ReconfigVM_Task, 'duration_secs': 1.278714} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.577346] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Reconfigured VM instance instance-0000003a to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1110.582824] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a76de31-e744-4e88-840c-9006c3086f2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.601870] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1110.601870] env[65121]: value = "task-5107260" [ 1110.601870] env[65121]: _type = "Task" [ 1110.601870] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.612776] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107260, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.629507] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107258, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.686421] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107259, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.857961] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.468s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.858350] env[65121]: WARNING neutronclient.v2_0.client [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.862134] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.151s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1110.863697] env[65121]: DEBUG nova.objects.instance [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1110.907216] env[65121]: INFO nova.network.neutron [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating port 04d47dd2-7f48-48e4-aed4-9d693168b8ca with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1110.964182] env[65121]: DEBUG oslo_concurrency.lockutils [req-bcfe4eac-4b1a-49b3-ae42-85720446010c req-40c41e05-4a8f-4f31-a621-beb26b85ba98 service nova] Releasing lock "refresh_cache-0497d253-887b-47a9-be8d-cc4592e046ba" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.060238] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.113981] env[65121]: DEBUG oslo_vmware.api [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107260, 'name': ReconfigVM_Task, 'duration_secs': 0.28597} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.114325] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993500', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'name': 'volume-02de1e93-4004-47b3-a1b4-d2690ef0da94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': 'e0d3e3c8-6471-4345-8677-369612674769', 'attached_at': '', 'detached_at': '', 'volume_id': '02de1e93-4004-47b3-a1b4-d2690ef0da94', 'serial': '02de1e93-4004-47b3-a1b4-d2690ef0da94'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1111.126499] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107258, 'name': CreateVM_Task, 'duration_secs': 0.776334} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.126745] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1111.127846] env[65121]: WARNING neutronclient.v2_0.client [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.128382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.128382] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.128685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1111.129087] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adca564e-817e-4fb2-9742-448f0cfc5279 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.134496] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1111.134496] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c6a2da-2e6d-91b6-f5c9-de0e966602ea" [ 1111.134496] env[65121]: _type = "Task" [ 1111.134496] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.142555] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c6a2da-2e6d-91b6-f5c9-de0e966602ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.181449] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107259, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605829} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.181711] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2/e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1111.181921] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.182206] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa04d35b-f1ed-400f-87c1-403d297c46d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.190127] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1111.190127] env[65121]: value = "task-5107261" [ 1111.190127] env[65121]: _type = "Task" [ 1111.190127] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.198759] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.575035] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "84ae6061-a962-4a74-8661-9718cc4c5346" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.575369] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.645739] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c6a2da-2e6d-91b6-f5c9-de0e966602ea, 'name': SearchDatastore_Task, 'duration_secs': 0.010423} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.645739] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.645962] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1111.646179] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.646343] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.646524] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1111.646801] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be8a8042-4042-42ae-aa91-2960b9988c46 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.661444] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1111.661674] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1111.663665] env[65121]: DEBUG nova.objects.instance [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'flavor' on Instance uuid e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.665056] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67392fc1-2d83-439d-b38c-c08fc63d9592 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.672512] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1111.672512] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e1fd27-e02d-2759-99a5-04926f24011f" [ 1111.672512] env[65121]: _type = "Task" [ 1111.672512] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.682550] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e1fd27-e02d-2759-99a5-04926f24011f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.703285] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095282} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.703571] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1111.704484] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a9c063-3827-4b3e-a45d-42a9e1f0260b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.730194] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2/e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.733171] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1c04d10-98d4-416c-a035-b037f8ddf43d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.754295] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1111.754295] env[65121]: value = "task-5107262" [ 1111.754295] env[65121]: _type = "Task" [ 1111.754295] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.766602] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107262, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.775180] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1111.775371] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1111.775609] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1111.775695] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1111.775863] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1111.776018] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1111.776264] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1111.776389] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1111.776562] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1111.776727] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1111.776910] env[65121]: DEBUG nova.virt.hardware [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1111.777833] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20a89b8-da4f-4155-a376-f3cc77282b59 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.787431] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52358c83-737a-47cd-adf5-09491be57963 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.802569] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:9a:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '53616727-ab6a-4f06-a597-0fc068a18955', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1111.810329] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1111.810675] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1111.810918] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e90f1c76-accd-4c0a-b355-35db6fecaa60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.833474] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1111.833474] env[65121]: value = "task-5107263" [ 1111.833474] env[65121]: _type = "Task" [ 1111.833474] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.842280] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107263, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.872753] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8d9db9bf-4766-470a-9dba-49ac47897008 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.873977] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.089s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.875506] env[65121]: INFO nova.compute.claims [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1111.978804] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquiring lock "4f86672c-425d-4641-a68c-2d6f5324e382" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.978972] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "4f86672c-425d-4641-a68c-2d6f5324e382" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.079115] env[65121]: DEBUG nova.compute.utils [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1112.185384] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e1fd27-e02d-2759-99a5-04926f24011f, 'name': SearchDatastore_Task, 'duration_secs': 0.031633} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.187129] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcf67784-8a93-45b0-9c34-bcb674ad0643 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.195115] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1112.195115] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525a1c08-67fc-614c-b7ff-8e444fd88c84" [ 1112.195115] env[65121]: _type = "Task" [ 1112.195115] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.204965] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525a1c08-67fc-614c-b7ff-8e444fd88c84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.265685] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107262, 'name': ReconfigVM_Task, 'duration_secs': 0.304942} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.266044] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Reconfigured VM instance instance-00000068 to attach disk [datastore1] e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2/e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.266824] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e91ae29-a9cf-4534-b4c8-62fbc56824ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.275230] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1112.275230] env[65121]: value = "task-5107264" [ 1112.275230] env[65121]: _type = "Task" [ 1112.275230] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.286898] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107264, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.327109] env[65121]: DEBUG nova.compute.manager [req-309c90ef-d0dd-4e26-84ff-4fe93243a009 req-f20437b7-8115-4b91-8f54-8f309f6639d4 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received event network-vif-plugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1112.327475] env[65121]: DEBUG oslo_concurrency.lockutils [req-309c90ef-d0dd-4e26-84ff-4fe93243a009 req-f20437b7-8115-4b91-8f54-8f309f6639d4 service nova] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.327847] env[65121]: DEBUG oslo_concurrency.lockutils [req-309c90ef-d0dd-4e26-84ff-4fe93243a009 req-f20437b7-8115-4b91-8f54-8f309f6639d4 service nova] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.328356] env[65121]: DEBUG oslo_concurrency.lockutils [req-309c90ef-d0dd-4e26-84ff-4fe93243a009 req-f20437b7-8115-4b91-8f54-8f309f6639d4 service nova] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.328740] env[65121]: DEBUG nova.compute.manager [req-309c90ef-d0dd-4e26-84ff-4fe93243a009 req-f20437b7-8115-4b91-8f54-8f309f6639d4 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] No waiting events found dispatching network-vif-plugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1112.328911] env[65121]: WARNING nova.compute.manager [req-309c90ef-d0dd-4e26-84ff-4fe93243a009 req-f20437b7-8115-4b91-8f54-8f309f6639d4 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received unexpected event network-vif-plugged-04d47dd2-7f48-48e4-aed4-9d693168b8ca for instance with vm_state shelved_offloaded and task_state spawning. [ 1112.345338] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107263, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.449480] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.449683] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.449844] env[65121]: DEBUG nova.network.neutron [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1112.481046] env[65121]: DEBUG nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1112.582038] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.673867] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0b1fff8-7a00-4a6a-b57d-5eb9bf1a88d6 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.307s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.706446] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525a1c08-67fc-614c-b7ff-8e444fd88c84, 'name': SearchDatastore_Task, 'duration_secs': 0.024045} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.706775] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1112.706946] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 0497d253-887b-47a9-be8d-cc4592e046ba/0497d253-887b-47a9-be8d-cc4592e046ba.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1112.707232] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d535b4e-9a23-496b-b63f-4d1bf1f7bc74 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.715495] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1112.715495] env[65121]: value = "task-5107265" [ 1112.715495] env[65121]: _type = "Task" [ 1112.715495] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.724552] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107265, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.785675] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107264, 'name': Rename_Task, 'duration_secs': 0.165791} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.785940] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1112.786286] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-882644d9-5f41-492d-a16b-07d4e20811e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.793715] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1112.793715] env[65121]: value = "task-5107266" [ 1112.793715] env[65121]: _type = "Task" [ 1112.793715] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.801905] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.844321] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107263, 'name': CreateVM_Task, 'duration_secs': 0.556646} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.844501] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1112.845134] env[65121]: WARNING neutronclient.v2_0.client [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1112.845530] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.845683] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.846013] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1112.846286] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6934c89-2670-4317-a895-f95546614eac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.851472] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1112.851472] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52578363-adb3-4d38-fdd1-8f850aee1278" [ 1112.851472] env[65121]: _type = "Task" [ 1112.851472] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.862409] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52578363-adb3-4d38-fdd1-8f850aee1278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.928854] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.929181] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.929385] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "e0d3e3c8-6471-4345-8677-369612674769-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.929561] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.929723] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.931950] env[65121]: INFO nova.compute.manager [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Terminating instance [ 1112.952108] env[65121]: WARNING neutronclient.v2_0.client [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.952929] env[65121]: WARNING openstack [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.953324] env[65121]: WARNING openstack [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.002741] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.069750] env[65121]: WARNING openstack [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.069750] env[65121]: WARNING openstack [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.148411] env[65121]: WARNING neutronclient.v2_0.client [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.149740] env[65121]: WARNING openstack [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.150259] env[65121]: WARNING openstack [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.233191] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df1914c-6627-4616-b7f5-72d43a1ff1db {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.241972] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107265, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.249233] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311e1d2b-a466-49e8-b732-1ac6cc86e6eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.292501] env[65121]: DEBUG nova.network.neutron [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1113.294817] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a91bc72-9b4d-47fd-982f-0321bcdf9b7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.315144] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107266, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.317455] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e9a213-f5f1-4559-994d-ec775203080d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.338503] env[65121]: DEBUG nova.compute.provider_tree [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.365111] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52578363-adb3-4d38-fdd1-8f850aee1278, 'name': SearchDatastore_Task, 'duration_secs': 0.010211} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.366129] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.366519] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1113.367153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.367232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.367747] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1113.368070] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b19ed9c2-f86c-444c-88f7-4caf7091e93f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.379428] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1113.379775] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1113.381047] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a64f7b1f-7a02-4767-994b-0b7f1655cf23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.388338] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1113.388338] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52262992-3484-9f62-8eee-b97c2e7106ae" [ 1113.388338] env[65121]: _type = "Task" [ 1113.388338] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.397106] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52262992-3484-9f62-8eee-b97c2e7106ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.436018] env[65121]: DEBUG nova.compute.manager [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1113.436270] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1113.437204] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a7cef9-3dea-4257-9cd2-2a90934ce0c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.446426] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1113.446737] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a2c8fb2-4ea2-4ac7-b52e-a4563ef086a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.454193] env[65121]: DEBUG oslo_vmware.api [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1113.454193] env[65121]: value = "task-5107267" [ 1113.454193] env[65121]: _type = "Task" [ 1113.454193] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.464233] env[65121]: DEBUG oslo_vmware.api [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107267, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.673540] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "84ae6061-a962-4a74-8661-9718cc4c5346" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.673872] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.674233] env[65121]: INFO nova.compute.manager [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Attaching volume 34ae2906-5e27-4132-9339-200c8a9a6ead to /dev/sdb [ 1113.710051] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40154afe-9d57-4472-a561-3b55dfc779ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.720191] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc069e3b-1a7b-4aa0-88b7-8c1905cf8742 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.727815] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107265, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607578} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.728078] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 0497d253-887b-47a9-be8d-cc4592e046ba/0497d253-887b-47a9-be8d-cc4592e046ba.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1113.728286] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1113.728518] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5194cdd2-4f12-4aa3-9beb-b003075dbf00 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.734575] env[65121]: DEBUG nova.virt.block_device [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating existing volume attachment record: 474f4b40-d21d-4365-bfd5-a5009dd22bf7 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1113.738150] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1113.738150] env[65121]: value = "task-5107268" [ 1113.738150] env[65121]: _type = "Task" [ 1113.738150] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.746516] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107268, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.799496] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.811467] env[65121]: DEBUG oslo_vmware.api [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107266, 'name': PowerOnVM_Task, 'duration_secs': 0.520951} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.811796] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1113.812065] env[65121]: INFO nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Took 8.37 seconds to spawn the instance on the hypervisor. [ 1113.812233] env[65121]: DEBUG nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1113.813089] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ceb6cb-f8fa-40e9-951b-d4a14f8115d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.831105] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='ad5feb5641cb92280cc3ccfdca052112',container_format='bare',created_at=2025-12-12T14:28:23Z,direct_url=,disk_format='vmdk',id=872a03ef-a1b4-487f-afd6-17a5c0d28e41,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-37570408-shelved',owner='48809b7bf36c4315b9ac315d6b9ca0c8',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2025-12-12T14:28:38Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1113.831346] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1113.831489] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1113.831692] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1113.832752] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1113.832752] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1113.832752] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.832752] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1113.832752] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1113.832752] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1113.833031] env[65121]: DEBUG nova.virt.hardware [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1113.834843] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffc96df-712e-4c2a-9cec-ae0c5b756ea9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.843751] env[65121]: DEBUG nova.scheduler.client.report [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1113.848525] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6890ba36-01b1-4700-8e68-ebe09216e5ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.864971] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:95:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04d47dd2-7f48-48e4-aed4-9d693168b8ca', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1113.873079] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1113.874536] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1113.874781] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98d34ad3-2c59-4d4a-bf36-718d042d42d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.901253] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52262992-3484-9f62-8eee-b97c2e7106ae, 'name': SearchDatastore_Task, 'duration_secs': 0.00986} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.903176] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1113.903176] env[65121]: value = "task-5107270" [ 1113.903176] env[65121]: _type = "Task" [ 1113.903176] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.903410] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-460a1a16-b266-46ea-a1a1-3c40f9c0680e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.912970] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1113.912970] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e302a1-5e2e-e478-1f1d-6f077b179fe2" [ 1113.912970] env[65121]: _type = "Task" [ 1113.912970] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.916077] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107270, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.925469] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e302a1-5e2e-e478-1f1d-6f077b179fe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.965761] env[65121]: DEBUG oslo_vmware.api [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107267, 'name': PowerOffVM_Task, 'duration_secs': 0.218634} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.966077] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1113.966277] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1113.966534] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-085d8afa-a126-45f9-bee1-7f6364c01ea1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.073330] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1114.073774] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1114.073986] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleting the datastore file [datastore1] e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.074304] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b29e280-903b-456f-9598-953aaa5f3dfe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.083081] env[65121]: DEBUG oslo_vmware.api [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1114.083081] env[65121]: value = "task-5107272" [ 1114.083081] env[65121]: _type = "Task" [ 1114.083081] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.092356] env[65121]: DEBUG oslo_vmware.api [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107272, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.249911] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107268, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06815} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.250420] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1114.251355] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95657e83-7ba6-4f17-99a2-d5c31e03956a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.279218] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 0497d253-887b-47a9-be8d-cc4592e046ba/0497d253-887b-47a9-be8d-cc4592e046ba.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.279706] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fafe9c0c-b7ec-4ddf-af1b-c6b0e909612c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.307238] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1114.307238] env[65121]: value = "task-5107275" [ 1114.307238] env[65121]: _type = "Task" [ 1114.307238] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.318534] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.334329] env[65121]: INFO nova.compute.manager [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Took 16.71 seconds to build instance. [ 1114.355092] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.356023] env[65121]: DEBUG nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1114.360763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 9.431s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.395362] env[65121]: DEBUG nova.compute.manager [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received event network-changed-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1114.395362] env[65121]: DEBUG nova.compute.manager [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Refreshing instance network info cache due to event network-changed-04d47dd2-7f48-48e4-aed4-9d693168b8ca. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1114.395362] env[65121]: DEBUG oslo_concurrency.lockutils [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Acquiring lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.396393] env[65121]: DEBUG oslo_concurrency.lockutils [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Acquired lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.398267] env[65121]: DEBUG nova.network.neutron [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Refreshing network info cache for port 04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1114.424999] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107270, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.432084] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e302a1-5e2e-e478-1f1d-6f077b179fe2, 'name': SearchDatastore_Task, 'duration_secs': 0.025364} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.432767] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1114.433135] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1114.433425] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a8b92eb-e90a-4751-adf4-2c4ade895b78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.440222] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "fc3109cb-72bb-4695-82c3-a323e619919d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1114.440458] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "fc3109cb-72bb-4695-82c3-a323e619919d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.440682] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "fc3109cb-72bb-4695-82c3-a323e619919d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1114.440858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "fc3109cb-72bb-4695-82c3-a323e619919d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.441027] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "fc3109cb-72bb-4695-82c3-a323e619919d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.444429] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1114.444429] env[65121]: value = "task-5107276" [ 1114.444429] env[65121]: _type = "Task" [ 1114.444429] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.444994] env[65121]: INFO nova.compute.manager [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Terminating instance [ 1114.458865] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107276, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.595081] env[65121]: DEBUG oslo_vmware.api [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107272, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.384227} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.595408] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1114.595609] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1114.595789] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1114.595969] env[65121]: INFO nova.compute.manager [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: e0d3e3c8-6471-4345-8677-369612674769] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1114.596296] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1114.596566] env[65121]: DEBUG nova.compute.manager [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1114.596702] env[65121]: DEBUG nova.network.neutron [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1114.597023] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.597605] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.597922] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.648896] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.823268] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107275, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.837659] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18dab7ce-8bdf-40d0-bfb5-a95871502237 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.231s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.868318] env[65121]: INFO nova.compute.claims [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1114.874761] env[65121]: DEBUG nova.compute.utils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1114.876668] env[65121]: DEBUG nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1114.876910] env[65121]: DEBUG nova.network.neutron [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1114.877418] env[65121]: WARNING neutronclient.v2_0.client [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.878221] env[65121]: WARNING neutronclient.v2_0.client [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.878581] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.879076] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.899955] env[65121]: WARNING neutronclient.v2_0.client [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.900783] env[65121]: WARNING openstack [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.901275] env[65121]: WARNING openstack [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.920334] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107270, 'name': CreateVM_Task, 'duration_secs': 0.658856} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.920925] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1114.921515] env[65121]: WARNING neutronclient.v2_0.client [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.921985] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.922192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.922619] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1114.922937] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dde10aaf-04e5-44d9-87fd-2d09fba7afe3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.930307] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1114.930307] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ea00fd-dc44-84ae-bac9-cf0a9efe2da5" [ 1114.930307] env[65121]: _type = "Task" [ 1114.930307] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.940365] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ea00fd-dc44-84ae-bac9-cf0a9efe2da5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.954097] env[65121]: DEBUG nova.policy [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1114.956530] env[65121]: DEBUG nova.compute.manager [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1114.956746] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1114.958321] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1db3814-e473-45a1-9888-ad8dd49f3e7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.964995] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107276, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.971731] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.972125] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbd0f958-506b-4b63-baae-e16636048476 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.981939] env[65121]: DEBUG oslo_vmware.api [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1114.981939] env[65121]: value = "task-5107277" [ 1114.981939] env[65121]: _type = "Task" [ 1114.981939] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.999759] env[65121]: DEBUG oslo_vmware.api [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.141091] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8781a2d0-5aa1-453a-9d97-e7bdcc401bf6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.148420] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6ada8ad9-5c1c-4383-ad4f-8e2154778e55 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Suspending the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1115.148594] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-c28f40f0-5481-4822-bd95-266806cca221 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.155775] env[65121]: DEBUG oslo_vmware.api [None req-6ada8ad9-5c1c-4383-ad4f-8e2154778e55 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1115.155775] env[65121]: value = "task-5107278" [ 1115.155775] env[65121]: _type = "Task" [ 1115.155775] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.165308] env[65121]: DEBUG oslo_vmware.api [None req-6ada8ad9-5c1c-4383-ad4f-8e2154778e55 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107278, 'name': SuspendVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.196148] env[65121]: DEBUG nova.compute.manager [req-e8e27076-9b15-4b9e-8bd4-1733ee1f8910 req-71d7de3f-749a-4041-b386-cf065d5e5d92 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Received event network-vif-deleted-71e0942f-5026-4128-ba81-16311feb9b3e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1115.196304] env[65121]: INFO nova.compute.manager [req-e8e27076-9b15-4b9e-8bd4-1733ee1f8910 req-71d7de3f-749a-4041-b386-cf065d5e5d92 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Neutron deleted interface 71e0942f-5026-4128-ba81-16311feb9b3e; detaching it from the instance and deleting it from the info cache [ 1115.196545] env[65121]: DEBUG nova.network.neutron [req-e8e27076-9b15-4b9e-8bd4-1733ee1f8910 req-71d7de3f-749a-4041-b386-cf065d5e5d92 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.220744] env[65121]: WARNING openstack [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.221248] env[65121]: WARNING openstack [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.320540] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107275, 'name': ReconfigVM_Task, 'duration_secs': 0.531186} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.320862] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 0497d253-887b-47a9-be8d-cc4592e046ba/0497d253-887b-47a9-be8d-cc4592e046ba.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.321577] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64d6dbda-8cb6-4efd-bb41-26acea22c9b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.333408] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1115.333408] env[65121]: value = "task-5107279" [ 1115.333408] env[65121]: _type = "Task" [ 1115.333408] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.343798] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107279, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.360402] env[65121]: WARNING neutronclient.v2_0.client [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1115.361293] env[65121]: WARNING openstack [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1115.361647] env[65121]: WARNING openstack [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1115.379097] env[65121]: INFO nova.compute.resource_tracker [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating resource usage from migration f186477a-a599-4c02-99d7-f463c3b63131 [ 1115.387595] env[65121]: DEBUG nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1115.454907] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.455237] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Processing image 872a03ef-a1b4-487f-afd6-17a5c0d28e41 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1115.455459] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41/872a03ef-a1b4-487f-afd6-17a5c0d28e41.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.455600] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "[datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41/872a03ef-a1b4-487f-afd6-17a5c0d28e41.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.455797] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1115.456439] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f23fc2f-ef2b-46bc-b0c7-225f52edcad1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.465879] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107276, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606129} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.469239] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1115.469699] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1115.470547] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ab6cab3-b77a-458f-b417-b6215462e0dd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.473372] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1115.473372] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1115.473808] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d39836d7-6af7-4fef-9454-5f2590e5d14c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.481723] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1115.481723] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5281a2d0-c3d4-b678-174b-dca602939778" [ 1115.481723] env[65121]: _type = "Task" [ 1115.481723] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.484246] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1115.484246] env[65121]: value = "task-5107280" [ 1115.484246] env[65121]: _type = "Task" [ 1115.484246] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.494563] env[65121]: DEBUG nova.network.neutron [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updated VIF entry in instance network info cache for port 04d47dd2-7f48-48e4-aed4-9d693168b8ca. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1115.495020] env[65121]: DEBUG nova.network.neutron [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.516949] env[65121]: DEBUG oslo_vmware.api [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107277, 'name': PowerOffVM_Task, 'duration_secs': 0.28642} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.518566] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107280, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.521902] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Preparing fetch location {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1115.522302] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Fetch image to [datastore2] OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68/OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68.vmdk {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1115.522550] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Downloading stream optimized image 872a03ef-a1b4-487f-afd6-17a5c0d28e41 to [datastore2] OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68/OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68.vmdk on the data store datastore2 as vApp {{(pid=65121) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1115.522806] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Downloading image file data 872a03ef-a1b4-487f-afd6-17a5c0d28e41 to the ESX as VM named 'OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68' {{(pid=65121) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1115.525456] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1115.525723] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1115.527222] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-628701d2-fcc3-448c-a0c0-253fe17e1e00 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.561510] env[65121]: DEBUG nova.network.neutron [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Successfully created port: ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1115.641043] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1115.641043] env[65121]: value = "resgroup-9" [ 1115.641043] env[65121]: _type = "ResourcePool" [ 1115.641043] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1115.641307] env[65121]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-154c4811-83d2-423f-93e1-9e27f12594c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.667054] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1115.667054] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1115.667054] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleting the datastore file [datastore1] fc3109cb-72bb-4695-82c3-a323e619919d {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.670884] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0858446d-0d8c-4487-b396-ddc1da4d8a6d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.673040] env[65121]: DEBUG nova.network.neutron [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.676495] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lease: (returnval){ [ 1115.676495] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d458f7-7c1b-5435-ce69-bb3c14854c68" [ 1115.676495] env[65121]: _type = "HttpNfcLease" [ 1115.676495] env[65121]: } obtained for vApp import into resource pool (val){ [ 1115.676495] env[65121]: value = "resgroup-9" [ 1115.676495] env[65121]: _type = "ResourcePool" [ 1115.676495] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1115.676832] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the lease: (returnval){ [ 1115.676832] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d458f7-7c1b-5435-ce69-bb3c14854c68" [ 1115.676832] env[65121]: _type = "HttpNfcLease" [ 1115.676832] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1115.680192] env[65121]: DEBUG oslo_vmware.api [None req-6ada8ad9-5c1c-4383-ad4f-8e2154778e55 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107278, 'name': SuspendVM_Task} progress is 62%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.684551] env[65121]: DEBUG oslo_vmware.api [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1115.684551] env[65121]: value = "task-5107283" [ 1115.684551] env[65121]: _type = "Task" [ 1115.684551] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.693076] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1115.693076] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d458f7-7c1b-5435-ce69-bb3c14854c68" [ 1115.693076] env[65121]: _type = "HttpNfcLease" [ 1115.693076] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1115.699164] env[65121]: DEBUG oslo_vmware.api [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.699714] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eaa173c4-6db7-4e1d-b6cf-5fd40bdc12fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.710232] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc89a51-0198-4440-8afb-ef9d52ef1e4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.751908] env[65121]: DEBUG nova.compute.manager [req-e8e27076-9b15-4b9e-8bd4-1733ee1f8910 req-71d7de3f-749a-4041-b386-cf065d5e5d92 service nova] [instance: e0d3e3c8-6471-4345-8677-369612674769] Detach interface failed, port_id=71e0942f-5026-4128-ba81-16311feb9b3e, reason: Instance e0d3e3c8-6471-4345-8677-369612674769 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1115.845120] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107279, 'name': Rename_Task, 'duration_secs': 0.24854} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.845496] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1115.845750] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43f7e9d5-f337-494b-8c01-036398435d35 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.849207] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeb5309-1ef1-438d-93cd-29af8e9bdb1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.859322] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3052b186-965c-4355-b69c-671b8ea08228 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.863243] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1115.863243] env[65121]: value = "task-5107284" [ 1115.863243] env[65121]: _type = "Task" [ 1115.863243] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.905181] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4570782a-3e4f-465f-b3e5-c5ca5b11e42f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.913327] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107284, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.920635] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34722d0-edbb-4fd7-bdcc-a97b80b7c983 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.933853] env[65121]: DEBUG nova.compute.provider_tree [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.003456] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107280, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079971} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.003456] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1116.004113] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ff9e85-b6b9-4048-8556-6b40d977ab1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.007994] env[65121]: DEBUG oslo_concurrency.lockutils [req-d424d176-a061-4822-b98c-747acc6a6ce0 req-99659b3a-f60e-4e86-933c-d3479e7659a1 service nova] Releasing lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.036346] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1116.036687] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48b36c52-126d-4004-8b7d-f8d6823ae853 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.059784] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1116.059784] env[65121]: value = "task-5107285" [ 1116.059784] env[65121]: _type = "Task" [ 1116.059784] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.072298] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107285, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.177252] env[65121]: INFO nova.compute.manager [-] [instance: e0d3e3c8-6471-4345-8677-369612674769] Took 1.58 seconds to deallocate network for instance. [ 1116.177741] env[65121]: DEBUG oslo_vmware.api [None req-6ada8ad9-5c1c-4383-ad4f-8e2154778e55 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107278, 'name': SuspendVM_Task, 'duration_secs': 0.798879} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.179367] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6ada8ad9-5c1c-4383-ad4f-8e2154778e55 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Suspended the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1116.179558] env[65121]: DEBUG nova.compute.manager [None req-6ada8ad9-5c1c-4383-ad4f-8e2154778e55 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1116.182574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4d49ff-dcb5-4cd6-8826-c79e14f866fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.202124] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.202124] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d458f7-7c1b-5435-ce69-bb3c14854c68" [ 1116.202124] env[65121]: _type = "HttpNfcLease" [ 1116.202124] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1116.202450] env[65121]: DEBUG oslo_vmware.api [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170189} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.202916] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1116.203109] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1116.203386] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1116.203570] env[65121]: INFO nova.compute.manager [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Took 1.25 seconds to destroy the instance on the hypervisor. [ 1116.203810] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1116.203996] env[65121]: DEBUG nova.compute.manager [-] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1116.204105] env[65121]: DEBUG nova.network.neutron [-] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1116.204596] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.204887] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.205151] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.284803] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.374663] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107284, 'name': PowerOnVM_Task} progress is 71%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.414022] env[65121]: DEBUG nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1116.437820] env[65121]: DEBUG nova.scheduler.client.report [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1116.453096] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1116.453367] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1116.454361] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1116.454592] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1116.454848] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1116.455016] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1116.455276] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1116.455418] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1116.455617] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1116.455793] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1116.455966] env[65121]: DEBUG nova.virt.hardware [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1116.456937] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277e9805-92f5-4787-aca4-3cebc975b0fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.467472] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64be139f-1bfe-4051-b29d-09eec06d8a30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.571054] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107285, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.688145] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.696105] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1116.696105] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d458f7-7c1b-5435-ce69-bb3c14854c68" [ 1116.696105] env[65121]: _type = "HttpNfcLease" [ 1116.696105] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1116.696600] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1116.696600] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d458f7-7c1b-5435-ce69-bb3c14854c68" [ 1116.696600] env[65121]: _type = "HttpNfcLease" [ 1116.696600] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1116.697508] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e61bb7-9f2d-481f-8104-d273887dce04 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.709062] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279004a-06bd-9187-6f67-9105426f021e/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1116.709062] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279004a-06bd-9187-6f67-9105426f021e/disk-0.vmdk. {{(pid=65121) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1116.773109] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a5811b72-a68f-42e8-b9d2-ba921f17a976 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.878216] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107284, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.946297] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.585s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.946558] env[65121]: INFO nova.compute.manager [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Migrating [ 1116.955364] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.789s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1116.957083] env[65121]: INFO nova.compute.claims [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1117.072721] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107285, 'name': ReconfigVM_Task, 'duration_secs': 0.825669} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.074292] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55/1fbd3a12-82d3-458d-a590-0f5c0c70ae55.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1117.075639] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'boot_index': 0, 'encrypted': False, 'encryption_format': None, 'encryption_options': None, 'device_type': 'disk', 'disk_bus': None, 'encryption_secret_uuid': None, 'guest_format': None, 'size': 0, 'image_id': '3ed5d656-a616-4609-8bb7-b02beb3ac3df'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'attachment_id': '166c385e-823d-4e12-bd55-d0b6f8b4633e', 'device_type': None, 'delete_on_termination': False, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'}, 'mount_device': '/dev/sdb', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65121) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1117.075934] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1117.076174] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1117.077564] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d102c8-c306-4588-bcb6-0d396f342d7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.100302] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2fe132-4409-4820-bb45-bd328b6494be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.135114] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] volume-41abb581-3841-4c0e-b79e-9820e1b241f1/volume-41abb581-3841-4c0e-b79e-9820e1b241f1.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.142997] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b18af806-118f-4652-ac05-007e2c322d3c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.161160] env[65121]: DEBUG nova.compute.manager [req-3d90e582-88ae-42c7-a258-875a6a40f010 req-88ca5784-01a1-4ab4-968d-1393a032e733 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Received event network-vif-deleted-65c052c9-c53e-40b1-a26b-6b8e80093cee {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1117.161380] env[65121]: INFO nova.compute.manager [req-3d90e582-88ae-42c7-a258-875a6a40f010 req-88ca5784-01a1-4ab4-968d-1393a032e733 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Neutron deleted interface 65c052c9-c53e-40b1-a26b-6b8e80093cee; detaching it from the instance and deleting it from the info cache [ 1117.161661] env[65121]: DEBUG nova.network.neutron [req-3d90e582-88ae-42c7-a258-875a6a40f010 req-88ca5784-01a1-4ab4-968d-1393a032e733 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1117.180014] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1117.180014] env[65121]: value = "task-5107287" [ 1117.180014] env[65121]: _type = "Task" [ 1117.180014] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.191452] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107287, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.209665] env[65121]: DEBUG nova.network.neutron [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Successfully updated port: ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1117.317412] env[65121]: DEBUG nova.network.neutron [-] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1117.380358] env[65121]: DEBUG oslo_vmware.api [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107284, 'name': PowerOnVM_Task, 'duration_secs': 1.223641} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.381415] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1117.381633] env[65121]: INFO nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Took 9.97 seconds to spawn the instance on the hypervisor. [ 1117.381812] env[65121]: DEBUG nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1117.382715] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804e518f-5e96-4cfb-aba0-ee892fbd1a70 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.457162] env[65121]: DEBUG nova.compute.manager [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-vif-plugged-ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1117.457162] env[65121]: DEBUG oslo_concurrency.lockutils [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Acquiring lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.457162] env[65121]: DEBUG oslo_concurrency.lockutils [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.457424] env[65121]: DEBUG oslo_concurrency.lockutils [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.457683] env[65121]: DEBUG nova.compute.manager [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] No waiting events found dispatching network-vif-plugged-ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1117.458291] env[65121]: WARNING nova.compute.manager [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received unexpected event network-vif-plugged-ff1a1ddf-0705-4111-80a1-faa79fe7c947 for instance with vm_state building and task_state spawning. [ 1117.458291] env[65121]: DEBUG nova.compute.manager [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1117.458601] env[65121]: DEBUG nova.compute.manager [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing instance network info cache due to event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1117.458682] env[65121]: DEBUG oslo_concurrency.lockutils [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.459065] env[65121]: DEBUG oslo_concurrency.lockutils [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.459330] env[65121]: DEBUG nova.network.neutron [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing network info cache for port ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1117.474557] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.474788] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.474960] env[65121]: DEBUG nova.network.neutron [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1117.671754] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ecbbe2f-2529-4c72-903b-b97782dd7ec6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.683025] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa63382a-d632-4a51-85b0-75dda17df109 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.709468] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107287, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.731277] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.731643] env[65121]: DEBUG nova.compute.manager [req-3d90e582-88ae-42c7-a258-875a6a40f010 req-88ca5784-01a1-4ab4-968d-1393a032e733 service nova] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Detach interface failed, port_id=65c052c9-c53e-40b1-a26b-6b8e80093cee, reason: Instance fc3109cb-72bb-4695-82c3-a323e619919d could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1117.820592] env[65121]: INFO nova.compute.manager [-] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Took 1.62 seconds to deallocate network for instance. [ 1117.904692] env[65121]: INFO nova.compute.manager [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Took 18.43 seconds to build instance. [ 1117.965503] env[65121]: WARNING neutronclient.v2_0.client [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.966031] env[65121]: WARNING openstack [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.966377] env[65121]: WARNING openstack [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.977037] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Completed reading data from the image iterator. {{(pid=65121) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1117.977037] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279004a-06bd-9187-6f67-9105426f021e/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1117.980401] env[65121]: WARNING neutronclient.v2_0.client [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1117.981044] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1117.981369] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1117.989110] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2152a42-2673-45f1-ad86-1a0d1024eb48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.998928] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279004a-06bd-9187-6f67-9105426f021e/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1117.998928] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279004a-06bd-9187-6f67-9105426f021e/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1117.999999] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-352805d7-87d9-4a52-b772-0cd40a8d698c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.007094] env[65121]: DEBUG nova.network.neutron [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1118.035549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.035969] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.036205] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.036400] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.036570] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1118.039973] env[65121]: INFO nova.compute.manager [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Terminating instance [ 1118.091250] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.091719] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.101883] env[65121]: DEBUG nova.network.neutron [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1118.177168] env[65121]: DEBUG oslo_vmware.rw_handles [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5279004a-06bd-9187-6f67-9105426f021e/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1118.177371] env[65121]: INFO nova.virt.vmwareapi.images [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Downloaded image file data 872a03ef-a1b4-487f-afd6-17a5c0d28e41 [ 1118.178520] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85415a2-841a-4aaa-80fa-99d2bd3c6155 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.182221] env[65121]: WARNING neutronclient.v2_0.client [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1118.182868] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.183178] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.212464] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a241e845-1d1d-4aaa-8b08-2ef7374d3e96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.220710] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107287, 'name': ReconfigVM_Task, 'duration_secs': 0.896286} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.221196] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfigured VM instance instance-0000005c to attach disk [datastore2] volume-41abb581-3841-4c0e-b79e-9820e1b241f1/volume-41abb581-3841-4c0e-b79e-9820e1b241f1.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.229324] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5067638b-c808-47ac-b89d-37146d98ab21 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.248254] env[65121]: INFO nova.virt.vmwareapi.images [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] The imported VM was unregistered [ 1118.251597] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Caching image {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1118.251901] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Creating directory with path [datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41 {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.253064] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3458bd8-2b92-414d-a2d0-dbf5beee97ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.261365] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1118.261365] env[65121]: value = "task-5107289" [ 1118.261365] env[65121]: _type = "Task" [ 1118.261365] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.272459] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107289, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.280583] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Created directory with path [datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41 {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.280778] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68/OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68.vmdk to [datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41/872a03ef-a1b4-487f-afd6-17a5c0d28e41.vmdk. {{(pid=65121) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1118.281059] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-abed6fba-dbb1-4a0a-bd34-1427241a0cea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.289936] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1118.289936] env[65121]: value = "task-5107290" [ 1118.289936] env[65121]: _type = "Task" [ 1118.289936] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.304796] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107290, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.314287] env[65121]: DEBUG nova.network.neutron [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance_info_cache with network_info: [{"id": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "address": "fa:16:3e:14:67:48", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3bafe6b-71", "ovs_interfaceid": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1118.329016] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.348731] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babb93d7-f8b4-43f8-aa0a-f4886462d37b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.358420] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd63cbd4-5856-45d1-b59c-6d9403100869 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.394472] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3d59a2-76bc-49d6-8f40-2005617f40ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.405449] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889ddbdc-dc48-4352-ac52-c8a63b952383 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.410215] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f4b1e8cc-f30e-4e6e-a07a-e9f70fec5371 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "0497d253-887b-47a9-be8d-cc4592e046ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.948s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1118.421632] env[65121]: DEBUG nova.compute.provider_tree [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.547030] env[65121]: DEBUG nova.compute.manager [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1118.547030] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1118.547614] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d9b97a-0b8a-4294-abf9-c729c4bc4587 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.556401] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.556699] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c53eea83-fb51-493f-a21c-7cbc96709d02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.606802] env[65121]: DEBUG oslo_concurrency.lockutils [req-11b46042-2630-4c5d-8408-13a21b8ffcc7 req-4a8fa80c-e7eb-49bb-aef5-64e5ea62ec13 service nova] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.607703] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1118.607703] env[65121]: DEBUG nova.network.neutron [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1118.622664] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.622723] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.622886] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleting the datastore file [datastore1] e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.623224] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e86299d6-6d3b-4a20-a92c-0bd0b166ef01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.630972] env[65121]: DEBUG oslo_vmware.api [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1118.630972] env[65121]: value = "task-5107292" [ 1118.630972] env[65121]: _type = "Task" [ 1118.630972] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.641686] env[65121]: DEBUG oslo_vmware.api [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.772678] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107289, 'name': ReconfigVM_Task, 'duration_secs': 0.260896} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.774028] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1118.774711] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-279be4fa-6de1-4b25-9c49-bcee14aed3cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.782597] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1118.782597] env[65121]: value = "task-5107293" [ 1118.782597] env[65121]: _type = "Task" [ 1118.782597] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.788966] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1118.789270] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993563', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'name': 'volume-34ae2906-5e27-4132-9339-200c8a9a6ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '84ae6061-a962-4a74-8661-9718cc4c5346', 'attached_at': '', 'detached_at': '', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'serial': '34ae2906-5e27-4132-9339-200c8a9a6ead'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1118.790493] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48528cbf-92e1-4d8f-aeb9-936f7a2045d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.802230] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107293, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.815300] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cff1293-27f7-4536-9cca-f84348f8ed8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.821730] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.823759] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107290, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.851118] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-34ae2906-5e27-4132-9339-200c8a9a6ead/volume-34ae2906-5e27-4132-9339-200c8a9a6ead.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.852225] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a793cd2d-538f-4477-bf5d-c6bf16120df4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.872748] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1118.872748] env[65121]: value = "task-5107294" [ 1118.872748] env[65121]: _type = "Task" [ 1118.872748] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.881954] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107294, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.926042] env[65121]: DEBUG nova.scheduler.client.report [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1119.112848] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.112848] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.142347] env[65121]: DEBUG oslo_vmware.api [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.154787] env[65121]: DEBUG nova.network.neutron [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1119.178375] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.178760] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.244952] env[65121]: WARNING neutronclient.v2_0.client [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.245839] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.246244] env[65121]: WARNING openstack [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.259612] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "0497d253-887b-47a9-be8d-cc4592e046ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.259879] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "0497d253-887b-47a9-be8d-cc4592e046ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.260114] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "0497d253-887b-47a9-be8d-cc4592e046ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.260291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "0497d253-887b-47a9-be8d-cc4592e046ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.260451] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "0497d253-887b-47a9-be8d-cc4592e046ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.263196] env[65121]: INFO nova.compute.manager [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Terminating instance [ 1119.295862] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107293, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.305578] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107290, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.375154] env[65121]: DEBUG nova.network.neutron [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1119.387737] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107294, 'name': ReconfigVM_Task, 'duration_secs': 0.417908} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.388070] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-34ae2906-5e27-4132-9339-200c8a9a6ead/volume-34ae2906-5e27-4132-9339-200c8a9a6ead.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.393587] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c304e75c-cafd-414a-ba3a-768210564a80 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.412311] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1119.412311] env[65121]: value = "task-5107295" [ 1119.412311] env[65121]: _type = "Task" [ 1119.412311] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.423332] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.432443] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.433237] env[65121]: DEBUG nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1119.437108] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.155s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.437478] env[65121]: DEBUG nova.objects.instance [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lazy-loading 'resources' on Instance uuid 39ca7a89-c262-4169-91b1-92ea45744518 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.642894] env[65121]: DEBUG oslo_vmware.api [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107292, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.954305} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.643233] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.643481] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.643674] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.643850] env[65121]: INFO nova.compute.manager [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1119.644169] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1119.644424] env[65121]: DEBUG nova.compute.manager [-] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1119.644527] env[65121]: DEBUG nova.network.neutron [-] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1119.644822] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.645430] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.645702] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.687681] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.768832] env[65121]: DEBUG nova.compute.manager [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1119.769175] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1119.770137] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590a4053-0d29-4776-a924-f9584f23e6c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.778702] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1119.779059] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e12e635-2f8a-4fbd-b174-76acae6f78ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.789058] env[65121]: DEBUG oslo_vmware.api [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1119.789058] env[65121]: value = "task-5107296" [ 1119.789058] env[65121]: _type = "Task" [ 1119.789058] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.797113] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107293, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.804283] env[65121]: DEBUG oslo_vmware.api [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.810228] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107290, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.879085] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1119.879255] env[65121]: DEBUG nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Instance network_info: |[{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1119.879657] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:f6:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9ec24851-7bb6-426b-b28f-f7b246df1713', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff1a1ddf-0705-4111-80a1-faa79fe7c947', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1119.888289] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1119.889319] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1119.889319] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b936e44-7ce8-466d-8091-f9681640f77b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.913796] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1119.913796] env[65121]: value = "task-5107297" [ 1119.913796] env[65121]: _type = "Task" [ 1119.913796] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.935579] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107295, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.936630] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107297, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.940749] env[65121]: DEBUG nova.compute.utils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1119.946858] env[65121]: DEBUG nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1119.947188] env[65121]: DEBUG nova.network.neutron [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1119.947511] env[65121]: WARNING neutronclient.v2_0.client [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.949990] env[65121]: WARNING neutronclient.v2_0.client [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.949990] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.949990] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.959043] env[65121]: DEBUG nova.compute.manager [req-8fd44286-3069-4458-9e59-256b3bf0aad3 req-1e5dbaf7-faa4-4298-8d04-b70fa09095f4 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Received event network-vif-deleted-98d8432f-dc60-460d-96b2-2c83d89b9c78 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1119.960245] env[65121]: INFO nova.compute.manager [req-8fd44286-3069-4458-9e59-256b3bf0aad3 req-1e5dbaf7-faa4-4298-8d04-b70fa09095f4 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Neutron deleted interface 98d8432f-dc60-460d-96b2-2c83d89b9c78; detaching it from the instance and deleting it from the info cache [ 1119.960245] env[65121]: DEBUG nova.network.neutron [req-8fd44286-3069-4458-9e59-256b3bf0aad3 req-1e5dbaf7-faa4-4298-8d04-b70fa09095f4 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1120.014027] env[65121]: DEBUG nova.policy [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db08f5913c504d07bfcfb417699eba4e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '20d977ed0564461894b3774afbf96a3a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1120.256121] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b15418-9213-47b2-9029-ef7d59867f4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.266102] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35914ba5-d62d-4528-bca7-ef82db87cb01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.314138] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f5bb07-e3ee-4b05-bf31-51525ae12c67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.326695] env[65121]: DEBUG oslo_vmware.api [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.326925] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107293, 'name': Rename_Task, 'duration_secs': 1.155734} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.333120] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1120.333616] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107290, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.334012] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f6e174e-38f5-430d-9f1e-9c0b42ea5b92 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.337521] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4518a1fc-f452-473c-b5d4-205252c21697 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.348648] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93422c1d-7664-4172-8429-0d9d9fd11fef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.353554] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1120.353554] env[65121]: value = "task-5107298" [ 1120.353554] env[65121]: _type = "Task" [ 1120.353554] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.365483] env[65121]: DEBUG nova.compute.provider_tree [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.386800] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance '9f9118f1-c708-469f-839e-b3c4ef52f38f' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1120.394240] env[65121]: DEBUG nova.network.neutron [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Successfully created port: a08a0ecc-509c-49a0-923d-4f9caf3e193e {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1120.408620] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107298, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.429561] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107297, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.433387] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107295, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.437853] env[65121]: DEBUG nova.network.neutron [-] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1120.444763] env[65121]: DEBUG nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1120.466022] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-add5ddf8-390f-43fc-8ee0-2ab0f5110cf3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.478496] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091819f9-f1e3-48a4-9e1c-7327079779fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.523193] env[65121]: DEBUG nova.compute.manager [req-8fd44286-3069-4458-9e59-256b3bf0aad3 req-1e5dbaf7-faa4-4298-8d04-b70fa09095f4 service nova] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Detach interface failed, port_id=98d8432f-dc60-460d-96b2-2c83d89b9c78, reason: Instance e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1120.817759] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107290, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.821070] env[65121]: DEBUG oslo_vmware.api [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107296, 'name': PowerOffVM_Task, 'duration_secs': 0.633485} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.821341] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1120.821527] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1120.821760] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61a83485-3903-4e86-a321-69f3cc6a4694 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.875572] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107298, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.888051] env[65121]: DEBUG nova.scheduler.client.report [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1120.898828] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1120.899206] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f711323-f442-4451-a045-9602f56b7aea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.907544] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1120.907544] env[65121]: value = "task-5107300" [ 1120.907544] env[65121]: _type = "Task" [ 1120.907544] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.917926] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107300, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.930711] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107297, 'name': CreateVM_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.935043] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1120.935270] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1120.935453] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleting the datastore file [datastore1] 0497d253-887b-47a9-be8d-cc4592e046ba {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1120.935746] env[65121]: DEBUG oslo_vmware.api [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107295, 'name': ReconfigVM_Task, 'duration_secs': 1.166442} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.936050] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e0bb16a-60ab-4dfa-adde-4070424cf482 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.938187] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993563', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'name': 'volume-34ae2906-5e27-4132-9339-200c8a9a6ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '84ae6061-a962-4a74-8661-9718cc4c5346', 'attached_at': '', 'detached_at': '', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'serial': '34ae2906-5e27-4132-9339-200c8a9a6ead'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1120.940239] env[65121]: INFO nova.compute.manager [-] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Took 1.30 seconds to deallocate network for instance. [ 1120.949208] env[65121]: DEBUG oslo_vmware.api [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1120.949208] env[65121]: value = "task-5107301" [ 1120.949208] env[65121]: _type = "Task" [ 1120.949208] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.963853] env[65121]: DEBUG oslo_vmware.api [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.316306] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107290, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.994002} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.316564] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68/OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68.vmdk to [datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41/872a03ef-a1b4-487f-afd6-17a5c0d28e41.vmdk. [ 1121.316848] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Cleaning up location [datastore2] OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1121.317057] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_21f76f46-ade9-4230-af4f-d301df863e68 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1121.317390] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90d519ba-6150-48ff-8b7a-03e216f3a382 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.325445] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1121.325445] env[65121]: value = "task-5107302" [ 1121.325445] env[65121]: _type = "Task" [ 1121.325445] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.334492] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.377630] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107298, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.393855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.957s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.397329] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.797s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1121.397329] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.397329] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1121.397843] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.338s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1121.399492] env[65121]: INFO nova.compute.claims [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1121.403771] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e030ad-3dc3-4063-8eba-d1831fbe3c80 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.421368] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf48562-b0a8-4972-ba27-7c6ff0f2e02d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.425559] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107300, 'name': PowerOffVM_Task, 'duration_secs': 0.456954} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.426704] env[65121]: INFO nova.scheduler.client.report [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Deleted allocations for instance 39ca7a89-c262-4169-91b1-92ea45744518 [ 1121.427868] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1121.428064] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance '9f9118f1-c708-469f-839e-b3c4ef52f38f' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1121.457662] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.458452] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107297, 'name': CreateVM_Task, 'duration_secs': 1.501057} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.460025] env[65121]: DEBUG nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1121.465590] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8dd5aed-480f-4138-8bf4-b685126825a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.468487] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1121.472209] env[65121]: WARNING neutronclient.v2_0.client [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.472209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.472209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.472209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1121.472484] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75b93679-8227-44d6-a31f-0e57f175d234 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.480580] env[65121]: DEBUG oslo_vmware.api [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.447177} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.484031] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1121.484031] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1121.484031] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1121.484031] env[65121]: INFO nova.compute.manager [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Took 1.71 seconds to destroy the instance on the hypervisor. [ 1121.484031] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1121.484031] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74837f11-055f-4640-8368-ca92ab77ba0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.488951] env[65121]: DEBUG nova.compute.manager [-] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1121.489128] env[65121]: DEBUG nova.network.neutron [-] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1121.489327] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.489874] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.490259] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.496880] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1121.496880] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52cfc5a2-5adf-639a-ead0-3df198bc905a" [ 1121.496880] env[65121]: _type = "Task" [ 1121.496880] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.530281] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177773MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1121.530281] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.533645] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1121.534122] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1121.534122] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1121.534284] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1121.534427] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1121.534597] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1121.534764] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1121.534913] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1121.535093] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1121.535253] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1121.535418] env[65121]: DEBUG nova.virt.hardware [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1121.539218] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca148ac8-51fc-4c02-9cbe-9f88ef258670 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.548517] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52cfc5a2-5adf-639a-ead0-3df198bc905a, 'name': SearchDatastore_Task, 'duration_secs': 0.021843} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.548517] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1121.548517] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1121.548517] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.548517] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.548517] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1121.549072] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85799693-df55-4e92-825e-e6ce86a18644 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.554446] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47b5d2a-5136-463b-a02f-7b15c08e4a56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.569025] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1121.569025] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1121.569025] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d362eb77-9a05-409c-a8ab-6e1a92a7ffcd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.586479] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1121.586479] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]526105e7-6c9b-8ec3-0961-c7099cb836ca" [ 1121.586479] env[65121]: _type = "Task" [ 1121.586479] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.598263] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526105e7-6c9b-8ec3-0961-c7099cb836ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.608162] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.836592] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.062346} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.836650] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1121.836812] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "[datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41/872a03ef-a1b4-487f-afd6-17a5c0d28e41.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1121.837050] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41/872a03ef-a1b4-487f-afd6-17a5c0d28e41.vmdk to [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1121.837314] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0d71c13-2f3f-4440-b3c7-189d522c082d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.846579] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1121.846579] env[65121]: value = "task-5107303" [ 1121.846579] env[65121]: _type = "Task" [ 1121.846579] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.852559] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107303, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.876455] env[65121]: DEBUG oslo_vmware.api [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107298, 'name': PowerOnVM_Task, 'duration_secs': 1.291207} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.876718] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.877075] env[65121]: DEBUG nova.compute.manager [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1121.878300] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3534c9ce-4be7-4ded-a2a4-a2ef743eee36 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.941507] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5303e59f-f2c8-4377-a971-d4487aae3629 tempest-ServersListShow296Test-210745977 tempest-ServersListShow296Test-210745977-project-member] Lock "39ca7a89-c262-4169-91b1-92ea45744518" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.011s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.945038] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1121.945394] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1121.946242] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1121.946242] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1121.946957] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1121.946957] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1121.947313] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1121.947932] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1121.947932] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1121.948363] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1121.948738] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1121.957925] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-374d5995-e497-42ac-813d-6e90c5d4921f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.981292] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1121.981292] env[65121]: value = "task-5107304" [ 1121.981292] env[65121]: _type = "Task" [ 1121.981292] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.991580] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107304, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.048961] env[65121]: DEBUG nova.objects.instance [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'flavor' on Instance uuid 84ae6061-a962-4a74-8661-9718cc4c5346 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.059044] env[65121]: DEBUG nova.network.neutron [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Successfully updated port: a08a0ecc-509c-49a0-923d-4f9caf3e193e {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1122.099773] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]526105e7-6c9b-8ec3-0961-c7099cb836ca, 'name': SearchDatastore_Task, 'duration_secs': 0.016988} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.101030] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dde1abe0-252a-436b-91bd-c9fee20a84a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.107677] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1122.107677] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525f2874-42a9-63b2-7d8c-79d782d4a4dd" [ 1122.107677] env[65121]: _type = "Task" [ 1122.107677] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.123152] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525f2874-42a9-63b2-7d8c-79d782d4a4dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.130529] env[65121]: DEBUG nova.compute.manager [req-e8d4bcec-73be-42a0-a847-bf00abaaabe3 req-07dfce74-7df4-4ca2-9087-decb431b16fa service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Received event network-vif-plugged-a08a0ecc-509c-49a0-923d-4f9caf3e193e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1122.130871] env[65121]: DEBUG oslo_concurrency.lockutils [req-e8d4bcec-73be-42a0-a847-bf00abaaabe3 req-07dfce74-7df4-4ca2-9087-decb431b16fa service nova] Acquiring lock "714930d8-3ec9-4aa8-831a-a9f28c365545-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.131200] env[65121]: DEBUG oslo_concurrency.lockutils [req-e8d4bcec-73be-42a0-a847-bf00abaaabe3 req-07dfce74-7df4-4ca2-9087-decb431b16fa service nova] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.131486] env[65121]: DEBUG oslo_concurrency.lockutils [req-e8d4bcec-73be-42a0-a847-bf00abaaabe3 req-07dfce74-7df4-4ca2-9087-decb431b16fa service nova] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.131770] env[65121]: DEBUG nova.compute.manager [req-e8d4bcec-73be-42a0-a847-bf00abaaabe3 req-07dfce74-7df4-4ca2-9087-decb431b16fa service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] No waiting events found dispatching network-vif-plugged-a08a0ecc-509c-49a0-923d-4f9caf3e193e {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1122.132089] env[65121]: WARNING nova.compute.manager [req-e8d4bcec-73be-42a0-a847-bf00abaaabe3 req-07dfce74-7df4-4ca2-9087-decb431b16fa service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Received unexpected event network-vif-plugged-a08a0ecc-509c-49a0-923d-4f9caf3e193e for instance with vm_state building and task_state spawning. [ 1122.152296] env[65121]: DEBUG nova.compute.manager [req-a78de3eb-bb39-4ccf-9e2c-76eabd8d59ab req-e3700b7c-a562-4f49-abbc-2ee5abd606cb service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Received event network-vif-deleted-d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1122.152438] env[65121]: INFO nova.compute.manager [req-a78de3eb-bb39-4ccf-9e2c-76eabd8d59ab req-e3700b7c-a562-4f49-abbc-2ee5abd606cb service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Neutron deleted interface d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d; detaching it from the instance and deleting it from the info cache [ 1122.152611] env[65121]: DEBUG nova.network.neutron [req-a78de3eb-bb39-4ccf-9e2c-76eabd8d59ab req-e3700b7c-a562-4f49-abbc-2ee5abd606cb service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1122.355664] env[65121]: DEBUG nova.network.neutron [-] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1122.357161] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107303, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.399471] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.497313] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107304, 'name': ReconfigVM_Task, 'duration_secs': 0.369655} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.497713] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance '9f9118f1-c708-469f-839e-b3c4ef52f38f' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1122.556805] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c186d885-5a19-4ed5-aa61-f92411575b51 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.883s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.564322] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "refresh_cache-714930d8-3ec9-4aa8-831a-a9f28c365545" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.564322] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquired lock "refresh_cache-714930d8-3ec9-4aa8-831a-a9f28c365545" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1122.564530] env[65121]: DEBUG nova.network.neutron [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1122.620550] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525f2874-42a9-63b2-7d8c-79d782d4a4dd, 'name': SearchDatastore_Task, 'duration_secs': 0.053176} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.620832] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1122.621144] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 8fce40bb-69c1-4fbb-9b0b-c7aece54d179/8fce40bb-69c1-4fbb-9b0b-c7aece54d179.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1122.621475] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c3f7c15-fd25-4b89-9039-c4205b68dd49 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.632931] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1122.632931] env[65121]: value = "task-5107305" [ 1122.632931] env[65121]: _type = "Task" [ 1122.632931] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.639900] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.656488] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90787729-8ea2-4059-a319-348d83a268d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.668734] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc59850-b9d9-4fbc-bc2d-4d432273135e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.712570] env[65121]: DEBUG nova.compute.manager [req-a78de3eb-bb39-4ccf-9e2c-76eabd8d59ab req-e3700b7c-a562-4f49-abbc-2ee5abd606cb service nova] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Detach interface failed, port_id=d3da86fe-ad3b-4fbc-bd1f-ba9d63fc793d, reason: Instance 0497d253-887b-47a9-be8d-cc4592e046ba could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1122.858443] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107303, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.861030] env[65121]: INFO nova.compute.manager [-] [instance: 0497d253-887b-47a9-be8d-cc4592e046ba] Took 1.37 seconds to deallocate network for instance. [ 1122.957266] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7853c57f-ea54-4dff-8a2d-7c7871afdc14 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.966262] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1f9aed-ec76-4f0b-b28f-57307640a9b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.006786] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1123.007054] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1123.007209] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1123.007419] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1123.007560] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1123.007700] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1123.007916] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.008090] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1123.008260] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1123.008415] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1123.008578] env[65121]: DEBUG nova.virt.hardware [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1123.013934] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1123.014537] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b06b6816-59b0-4bff-a7cf-ba167e8f04cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.029382] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd572862-28f8-4268-9d55-f1be3cad36a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.042846] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3847a10-93c4-4cca-920a-496de33f214c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.047318] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1123.047318] env[65121]: value = "task-5107306" [ 1123.047318] env[65121]: _type = "Task" [ 1123.047318] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.060147] env[65121]: DEBUG nova.compute.provider_tree [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.065976] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107306, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.067093] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.067459] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.109335] env[65121]: DEBUG nova.network.neutron [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1123.135999] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.136663] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.152844] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.206629] env[65121]: WARNING neutronclient.v2_0.client [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1123.207189] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.207610] env[65121]: WARNING openstack [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.300254] env[65121]: DEBUG nova.compute.manager [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1123.332029] env[65121]: DEBUG nova.network.neutron [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Updating instance_info_cache with network_info: [{"id": "a08a0ecc-509c-49a0-923d-4f9caf3e193e", "address": "fa:16:3e:48:5e:a7", "network": {"id": "35296256-512f-4283-80b6-784a5addec1c", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-889700843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "20d977ed0564461894b3774afbf96a3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa08a0ecc-50", "ovs_interfaceid": "a08a0ecc-509c-49a0-923d-4f9caf3e193e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1123.362667] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107303, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.368601] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.561864] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.568124] env[65121]: DEBUG nova.scheduler.client.report [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1123.643915] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.825825] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.835278] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Releasing lock "refresh_cache-714930d8-3ec9-4aa8-831a-a9f28c365545" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1123.835725] env[65121]: DEBUG nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Instance network_info: |[{"id": "a08a0ecc-509c-49a0-923d-4f9caf3e193e", "address": "fa:16:3e:48:5e:a7", "network": {"id": "35296256-512f-4283-80b6-784a5addec1c", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-889700843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "20d977ed0564461894b3774afbf96a3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa08a0ecc-50", "ovs_interfaceid": "a08a0ecc-509c-49a0-923d-4f9caf3e193e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1123.836295] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:5e:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd903c404-a23a-40c0-a217-96d4bb2e5b08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a08a0ecc-509c-49a0-923d-4f9caf3e193e', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.844594] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Creating folder: Project (20d977ed0564461894b3774afbf96a3a). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1123.845040] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-370252c7-7b54-408e-98b5-35018b870e08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.859022] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107303, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.866238] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Created folder: Project (20d977ed0564461894b3774afbf96a3a) in parent group-v993268. [ 1123.866475] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Creating folder: Instances. Parent ref: group-v993566. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1123.866859] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6fb063da-0912-4798-b115-9868aaa74aa0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.883472] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Created folder: Instances in parent group-v993566. [ 1123.883777] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1123.884039] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1123.884281] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0ee09fc-49fd-4d9d-97cb-d3a3d2afe405 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.911063] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.911063] env[65121]: value = "task-5107309" [ 1123.911063] env[65121]: _type = "Task" [ 1123.911063] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.925143] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107309, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.059777] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107306, 'name': ReconfigVM_Task, 'duration_secs': 0.604033} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.060145] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1124.061023] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2ac90b-0c14-4cad-9aba-96e47b3b1d8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.080865] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.683s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.081418] env[65121]: DEBUG nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1124.091739] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 9f9118f1-c708-469f-839e-b3c4ef52f38f/9f9118f1-c708-469f-839e-b3c4ef52f38f.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.092273] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.090s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.093901] env[65121]: INFO nova.compute.claims [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.096774] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df1c3ca0-ea46-4783-b3f2-e6651dab56e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.120103] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1124.120103] env[65121]: value = "task-5107310" [ 1124.120103] env[65121]: _type = "Task" [ 1124.120103] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.131885] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107310, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.143832] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.162258] env[65121]: DEBUG nova.compute.manager [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Received event network-changed-a08a0ecc-509c-49a0-923d-4f9caf3e193e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1124.162331] env[65121]: DEBUG nova.compute.manager [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Refreshing instance network info cache due to event network-changed-a08a0ecc-509c-49a0-923d-4f9caf3e193e. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1124.162522] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Acquiring lock "refresh_cache-714930d8-3ec9-4aa8-831a-a9f28c365545" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.162655] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Acquired lock "refresh_cache-714930d8-3ec9-4aa8-831a-a9f28c365545" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.162810] env[65121]: DEBUG nova.network.neutron [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Refreshing network info cache for port a08a0ecc-509c-49a0-923d-4f9caf3e193e {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1124.363948] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107303, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.423857] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107309, 'name': CreateVM_Task, 'duration_secs': 0.393331} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.424041] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1124.424637] env[65121]: WARNING neutronclient.v2_0.client [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.425027] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.425223] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.425578] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1124.426326] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec0b1a6-f756-4dd7-98ae-01ca41ef0d26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.434677] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1124.434677] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52721cbb-e1e9-6c37-27fc-ee84fdf821eb" [ 1124.434677] env[65121]: _type = "Task" [ 1124.434677] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.445980] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52721cbb-e1e9-6c37-27fc-ee84fdf821eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.594317] env[65121]: DEBUG nova.compute.utils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1124.595982] env[65121]: DEBUG nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1124.596149] env[65121]: DEBUG nova.network.neutron [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1124.596854] env[65121]: WARNING neutronclient.v2_0.client [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.596854] env[65121]: WARNING neutronclient.v2_0.client [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.597464] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.597832] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.636659] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107310, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.648576] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107305, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.664041] env[65121]: DEBUG nova.policy [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af00bd582c0843949491bbcecbfcd2dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72237a0d762645588c41231b0a34a796', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1124.666828] env[65121]: WARNING neutronclient.v2_0.client [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.667610] env[65121]: WARNING openstack [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.668074] env[65121]: WARNING openstack [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.776245] env[65121]: WARNING openstack [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.776630] env[65121]: WARNING openstack [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.863361] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107303, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.899304} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.865072] env[65121]: WARNING neutronclient.v2_0.client [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.865939] env[65121]: WARNING openstack [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.866809] env[65121]: WARNING openstack [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.875464] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/872a03ef-a1b4-487f-afd6-17a5c0d28e41/872a03ef-a1b4-487f-afd6-17a5c0d28e41.vmdk to [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1124.876804] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ab6bb8-fc5b-464f-96bb-85f5f3e80e4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.902806] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.903570] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53a9ee15-fb5a-43ba-aa99-57f67503ebb0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.927265] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1124.927265] env[65121]: value = "task-5107311" [ 1124.927265] env[65121]: _type = "Task" [ 1124.927265] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.939049] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107311, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.951362] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52721cbb-e1e9-6c37-27fc-ee84fdf821eb, 'name': SearchDatastore_Task, 'duration_secs': 0.050224} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.951711] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.952031] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1124.952269] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.952438] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.952631] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1124.952955] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7eecf4dc-84f7-43ce-ae49-6667a8c75d00 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.963956] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1124.964233] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1124.965083] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08b2d5af-1dd2-4529-bbad-79d82ba049c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.969783] env[65121]: DEBUG nova.network.neutron [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Updated VIF entry in instance network info cache for port a08a0ecc-509c-49a0-923d-4f9caf3e193e. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1124.970090] env[65121]: DEBUG nova.network.neutron [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Updating instance_info_cache with network_info: [{"id": "a08a0ecc-509c-49a0-923d-4f9caf3e193e", "address": "fa:16:3e:48:5e:a7", "network": {"id": "35296256-512f-4283-80b6-784a5addec1c", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-889700843-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "20d977ed0564461894b3774afbf96a3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa08a0ecc-50", "ovs_interfaceid": "a08a0ecc-509c-49a0-923d-4f9caf3e193e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.975127] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1124.975127] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528c0125-d040-c651-376f-3c81a24843b0" [ 1124.975127] env[65121]: _type = "Task" [ 1124.975127] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.984501] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528c0125-d040-c651-376f-3c81a24843b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.016372] env[65121]: DEBUG nova.network.neutron [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Successfully created port: 99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1125.106832] env[65121]: DEBUG nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1125.131629] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107310, 'name': ReconfigVM_Task, 'duration_secs': 0.987625} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.131933] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 9f9118f1-c708-469f-839e-b3c4ef52f38f/9f9118f1-c708-469f-839e-b3c4ef52f38f.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.132296] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance '9f9118f1-c708-469f-839e-b3c4ef52f38f' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1125.150994] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107305, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.339171} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.153781] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 8fce40bb-69c1-4fbb-9b0b-c7aece54d179/8fce40bb-69c1-4fbb-9b0b-c7aece54d179.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1125.154028] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1125.155170] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8eaa2e04-5643-427d-9a21-f0d2018a3f9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.164596] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1125.164596] env[65121]: value = "task-5107312" [ 1125.164596] env[65121]: _type = "Task" [ 1125.164596] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.177569] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107312, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.408555] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bab5dc7-2f76-4886-a819-c73054e46779 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.417597] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45bbd1da-883d-455f-bf1f-854cc43e9184 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.453063] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a3e673-b465-4139-b6a6-fc251c5a2e2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.467856] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863e0420-2019-4758-b678-f27bee68fbb9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.471941] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.472501] env[65121]: DEBUG oslo_concurrency.lockutils [req-e7b62275-694c-4d10-90f1-26187baaf330 req-7f9c7aa6-eefe-4d9c-8d74-73e2748d20d6 service nova] Releasing lock "refresh_cache-714930d8-3ec9-4aa8-831a-a9f28c365545" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.483841] env[65121]: DEBUG nova.compute.provider_tree [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.495016] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528c0125-d040-c651-376f-3c81a24843b0, 'name': SearchDatastore_Task, 'duration_secs': 0.014224} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.496509] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e03d209-95cd-4587-a672-1cc58195c7c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.503638] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1125.503638] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52748d58-5be0-3362-aeba-229ca02ed351" [ 1125.503638] env[65121]: _type = "Task" [ 1125.503638] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.515427] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52748d58-5be0-3362-aeba-229ca02ed351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.646015] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe2f6ef-0844-4681-9557-748ecffc38e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.672267] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983a46eb-af36-4ced-ba85-6b488829cd18 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.694930] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107312, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073207} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.695626] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance '9f9118f1-c708-469f-839e-b3c4ef52f38f' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1125.699040] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1125.700133] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdca3d6d-d524-478f-9a54-d2dd3d0fb0fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.726653] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 8fce40bb-69c1-4fbb-9b0b-c7aece54d179/8fce40bb-69c1-4fbb-9b0b-c7aece54d179.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1125.727394] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e7b334c-64ed-492b-89e9-d2ea012bfe18 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.751604] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1125.751604] env[65121]: value = "task-5107313" [ 1125.751604] env[65121]: _type = "Task" [ 1125.751604] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.761780] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107313, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.960543] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107311, 'name': ReconfigVM_Task, 'duration_secs': 0.837603} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.960749] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731/52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.961483] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf7380a1-f751-48a0-ad41-ec5f1a711515 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.969570] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1125.969570] env[65121]: value = "task-5107314" [ 1125.969570] env[65121]: _type = "Task" [ 1125.969570] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.979093] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107314, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.990288] env[65121]: DEBUG nova.scheduler.client.report [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1126.015867] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52748d58-5be0-3362-aeba-229ca02ed351, 'name': SearchDatastore_Task, 'duration_secs': 0.072014} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.016190] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1126.016463] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 714930d8-3ec9-4aa8-831a-a9f28c365545/714930d8-3ec9-4aa8-831a-a9f28c365545.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1126.016786] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7adcb56c-ab73-4614-b9c1-92d5f2bea4c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.024736] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1126.024736] env[65121]: value = "task-5107315" [ 1126.024736] env[65121]: _type = "Task" [ 1126.024736] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.045229] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107315, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.116856] env[65121]: DEBUG nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1126.159882] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1126.160280] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1126.160539] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1126.161458] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1126.161458] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1126.161458] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1126.161775] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.162035] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1126.162333] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1126.162592] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1126.162875] env[65121]: DEBUG nova.virt.hardware [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1126.164211] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1babfb-12c4-4c6f-a5ff-3a0acabcf337 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.175312] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c5dc58-5364-457b-bfc0-f13913ac6d90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.202603] env[65121]: WARNING neutronclient.v2_0.client [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.262435] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107313, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.439815] env[65121]: DEBUG nova.network.neutron [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Port b3bafe6b-71bc-45d0-bf37-29da60b15de0 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1126.481386] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107314, 'name': Rename_Task, 'duration_secs': 0.464724} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.481683] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1126.481979] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d4f23d7-d935-46df-8708-83c15c5c34b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.490339] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1126.490339] env[65121]: value = "task-5107316" [ 1126.490339] env[65121]: _type = "Task" [ 1126.490339] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.495832] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.496259] env[65121]: DEBUG nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1126.502935] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.815s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.503286] env[65121]: DEBUG nova.objects.instance [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'resources' on Instance uuid e0d3e3c8-6471-4345-8677-369612674769 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.505345] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107316, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.537373] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107315, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.656485] env[65121]: DEBUG nova.compute.manager [req-cef89777-d843-4e29-8f18-f4c492b4f671 req-28e94882-d986-4c83-a20c-5f6c60a1f75c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Received event network-vif-plugged-99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1126.657054] env[65121]: DEBUG oslo_concurrency.lockutils [req-cef89777-d843-4e29-8f18-f4c492b4f671 req-28e94882-d986-4c83-a20c-5f6c60a1f75c service nova] Acquiring lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.657761] env[65121]: DEBUG oslo_concurrency.lockutils [req-cef89777-d843-4e29-8f18-f4c492b4f671 req-28e94882-d986-4c83-a20c-5f6c60a1f75c service nova] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.657975] env[65121]: DEBUG oslo_concurrency.lockutils [req-cef89777-d843-4e29-8f18-f4c492b4f671 req-28e94882-d986-4c83-a20c-5f6c60a1f75c service nova] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.658268] env[65121]: DEBUG nova.compute.manager [req-cef89777-d843-4e29-8f18-f4c492b4f671 req-28e94882-d986-4c83-a20c-5f6c60a1f75c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] No waiting events found dispatching network-vif-plugged-99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1126.658453] env[65121]: WARNING nova.compute.manager [req-cef89777-d843-4e29-8f18-f4c492b4f671 req-28e94882-d986-4c83-a20c-5f6c60a1f75c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Received unexpected event network-vif-plugged-99be1255-3e71-41e6-936a-acbf44bf9de9 for instance with vm_state building and task_state spawning. [ 1126.732508] env[65121]: DEBUG nova.network.neutron [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Successfully updated port: 99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1126.764715] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107313, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.002267] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107316, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.006819] env[65121]: DEBUG nova.compute.utils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1127.011300] env[65121]: DEBUG nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1127.011627] env[65121]: DEBUG nova.network.neutron [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1127.011739] env[65121]: WARNING neutronclient.v2_0.client [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.012032] env[65121]: WARNING neutronclient.v2_0.client [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.012655] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.012924] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.043254] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107315, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.895547} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.043583] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 714930d8-3ec9-4aa8-831a-a9f28c365545/714930d8-3ec9-4aa8-831a-a9f28c365545.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1127.043868] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1127.044202] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ba4e5a6-ddf0-4719-a697-37174c6b8b4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.056128] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1127.056128] env[65121]: value = "task-5107317" [ 1127.056128] env[65121]: _type = "Task" [ 1127.056128] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.063990] env[65121]: DEBUG nova.policy [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a342c96eac6b481dbc078c361b5b5b1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '529fe6e1d6754df5b4682544b4d3aaa1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1127.072221] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107317, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.235752] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.235947] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1127.236181] env[65121]: DEBUG nova.network.neutron [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1127.268656] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107313, 'name': ReconfigVM_Task, 'duration_secs': 1.015839} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.268788] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 8fce40bb-69c1-4fbb-9b0b-c7aece54d179/8fce40bb-69c1-4fbb-9b0b-c7aece54d179.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1127.269738] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca750252-ae2c-4e7c-9fbf-083faef4a1d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.276434] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1127.276434] env[65121]: value = "task-5107318" [ 1127.276434] env[65121]: _type = "Task" [ 1127.276434] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.286080] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107318, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.302171] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03314ed1-3f5c-42ba-b67a-ab93b27013ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.314441] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc46184-c831-480c-a177-5f1ed94e85ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.369871] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4561880d-e6a4-4fc7-b41f-906815b8a9c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.385528] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f14494-0d09-4c4e-b5f6-b0c6b251c6ea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.409980] env[65121]: DEBUG nova.compute.provider_tree [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.444866] env[65121]: DEBUG nova.network.neutron [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Successfully created port: 1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1127.458277] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.458510] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1127.458680] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1127.502766] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107316, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.511738] env[65121]: DEBUG nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1127.567673] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107317, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.344893} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.568036] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.568959] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db9901b-aab0-49f7-b940-29576665929e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.594838] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 714930d8-3ec9-4aa8-831a-a9f28c365545/714930d8-3ec9-4aa8-831a-a9f28c365545.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.595560] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38048506-e7e0-4488-920f-3128fcd28011 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.617419] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1127.617419] env[65121]: value = "task-5107319" [ 1127.617419] env[65121]: _type = "Task" [ 1127.617419] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.627630] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.739301] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.739892] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.789900] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107318, 'name': Rename_Task, 'duration_secs': 0.228825} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.790337] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1127.790685] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52830c83-8ca4-46bd-a45b-824657ff3a1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.798681] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1127.798681] env[65121]: value = "task-5107320" [ 1127.798681] env[65121]: _type = "Task" [ 1127.798681] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.810840] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107320, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.812195] env[65121]: DEBUG nova.network.neutron [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1127.864974] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.865581] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.915406] env[65121]: DEBUG nova.scheduler.client.report [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1127.977383] env[65121]: WARNING neutronclient.v2_0.client [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.978140] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.978486] env[65121]: WARNING openstack [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1128.004082] env[65121]: DEBUG oslo_vmware.api [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107316, 'name': PowerOnVM_Task, 'duration_secs': 1.136126} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.004828] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.020022] env[65121]: INFO nova.virt.block_device [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Booting with volume 20d5c722-822e-46f0-9b9b-7f7d92bd114a at /dev/sda [ 1128.059119] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9b52303-dab1-4f89-9cbf-762337966ddd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.074041] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da3e213-0e98-466b-82a2-0d7d18be11ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.117774] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d98716b-9970-4588-87a1-9a43564a085b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.129548] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107319, 'name': ReconfigVM_Task, 'duration_secs': 0.314898} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.131070] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 714930d8-3ec9-4aa8-831a-a9f28c365545/714930d8-3ec9-4aa8-831a-a9f28c365545.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.131747] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92c0fac6-0c32-41b5-9ed5-0d33a2f3c40a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.135938] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f66d32-ca0f-4256-ab3e-91f8f2032108 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.147338] env[65121]: DEBUG nova.compute.manager [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1128.149585] env[65121]: DEBUG nova.network.neutron [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1128.151249] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c5b0de-0666-4737-9ec5-3d465cdd23a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.167376] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1128.167376] env[65121]: value = "task-5107321" [ 1128.167376] env[65121]: _type = "Task" [ 1128.167376] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.183546] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd9e027-d7c2-446e-94e9-c7d907d3ed3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.192265] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107321, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.197054] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2f3054-76bd-4d89-b8c4-ad6c5316a5b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.211559] env[65121]: DEBUG nova.virt.block_device [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updating existing volume attachment record: 219e8d7a-c5fa-4b8f-a91a-e971ee2b447d {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1128.310618] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107320, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.427697] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.925s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.430146] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.101s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.430400] env[65121]: DEBUG nova.objects.instance [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'resources' on Instance uuid fc3109cb-72bb-4695-82c3-a323e619919d {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.452890] env[65121]: INFO nova.scheduler.client.report [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted allocations for instance e0d3e3c8-6471-4345-8677-369612674769 [ 1128.471045] env[65121]: WARNING neutronclient.v2_0.client [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1128.505774] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.505774] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1128.506045] env[65121]: DEBUG nova.network.neutron [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1128.655646] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1128.656120] env[65121]: DEBUG nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Instance network_info: |[{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1128.656667] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:a0:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6966f473-59ac-49bb-9b7a-22c61f4e61e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99be1255-3e71-41e6-936a-acbf44bf9de9', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1128.665801] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1128.669973] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1128.671725] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98ab1ed0-0670-410e-aecb-65059ad91d37 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.691822] env[65121]: DEBUG oslo_concurrency.lockutils [None req-dc2ea705-394a-47d1-9731-420ce8b07645 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 27.169s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.700458] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107321, 'name': Rename_Task, 'duration_secs': 0.150442} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.703136] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1128.704305] env[65121]: DEBUG nova.compute.manager [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Received event network-changed-99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1128.704494] env[65121]: DEBUG nova.compute.manager [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Refreshing instance network info cache due to event network-changed-99be1255-3e71-41e6-936a-acbf44bf9de9. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1128.704762] env[65121]: DEBUG oslo_concurrency.lockutils [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Acquiring lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.704904] env[65121]: DEBUG oslo_concurrency.lockutils [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Acquired lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1128.705076] env[65121]: DEBUG nova.network.neutron [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Refreshing network info cache for port 99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1128.706723] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1128.706723] env[65121]: value = "task-5107322" [ 1128.706723] env[65121]: _type = "Task" [ 1128.706723] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.707135] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49324f70-9751-4642-83cf-901efcd52b1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.721843] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107322, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.723593] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1128.723593] env[65121]: value = "task-5107323" [ 1128.723593] env[65121]: _type = "Task" [ 1128.723593] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.733683] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.814640] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107320, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.966708] env[65121]: DEBUG nova.compute.manager [req-912ff8e5-44ef-4ce6-bfc4-d7b200e1f0c8 req-8061f1c8-f610-4024-95d8-5fb20b86286c service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Received event network-vif-plugged-1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1128.966978] env[65121]: DEBUG oslo_concurrency.lockutils [req-912ff8e5-44ef-4ce6-bfc4-d7b200e1f0c8 req-8061f1c8-f610-4024-95d8-5fb20b86286c service nova] Acquiring lock "4f86672c-425d-4641-a68c-2d6f5324e382-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.967268] env[65121]: DEBUG oslo_concurrency.lockutils [req-912ff8e5-44ef-4ce6-bfc4-d7b200e1f0c8 req-8061f1c8-f610-4024-95d8-5fb20b86286c service nova] Lock "4f86672c-425d-4641-a68c-2d6f5324e382-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.967434] env[65121]: DEBUG oslo_concurrency.lockutils [req-912ff8e5-44ef-4ce6-bfc4-d7b200e1f0c8 req-8061f1c8-f610-4024-95d8-5fb20b86286c service nova] Lock "4f86672c-425d-4641-a68c-2d6f5324e382-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.967600] env[65121]: DEBUG nova.compute.manager [req-912ff8e5-44ef-4ce6-bfc4-d7b200e1f0c8 req-8061f1c8-f610-4024-95d8-5fb20b86286c service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] No waiting events found dispatching network-vif-plugged-1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1128.967753] env[65121]: WARNING nova.compute.manager [req-912ff8e5-44ef-4ce6-bfc4-d7b200e1f0c8 req-8061f1c8-f610-4024-95d8-5fb20b86286c service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Received unexpected event network-vif-plugged-1d9c1075-163b-42c1-b44d-49a4699c9930 for instance with vm_state building and task_state block_device_mapping. [ 1128.968590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-62312f15-80d6-4335-be4f-76ca67b1f802 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "e0d3e3c8-6471-4345-8677-369612674769" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.039s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.009464] env[65121]: WARNING neutronclient.v2_0.client [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.010158] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.010511] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.070506] env[65121]: DEBUG nova.network.neutron [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Successfully updated port: 1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1129.141249] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.141637] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.211166] env[65121]: WARNING neutronclient.v2_0.client [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.211959] env[65121]: WARNING openstack [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.212351] env[65121]: WARNING openstack [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.222156] env[65121]: WARNING neutronclient.v2_0.client [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.222841] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.224131] env[65121]: WARNING openstack [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.250442] env[65121]: DEBUG oslo_vmware.api [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107323, 'name': PowerOnVM_Task, 'duration_secs': 0.512338} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.254268] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1129.254268] env[65121]: INFO nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Took 7.79 seconds to spawn the instance on the hypervisor. [ 1129.254472] env[65121]: DEBUG nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1129.255060] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107322, 'name': CreateVM_Task, 'duration_secs': 0.390386} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.257852] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0419848-b945-49cd-8bd4-bed4238e50f1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.260373] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1129.262025] env[65121]: WARNING neutronclient.v2_0.client [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.263096] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.263096] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.263096] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1129.263594] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a47fcd-4f76-4e42-9252-c05187b50fbc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.266578] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d81f32c1-ee68-4ec0-9886-d5501d6d6e2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.280040] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1129.280040] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab928c-d375-2c45-8da7-36a1c17bf0ab" [ 1129.280040] env[65121]: _type = "Task" [ 1129.280040] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.281167] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bad7ccd-b38e-4bb5-81a1-cb3c2256a867 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.299630] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ab928c-d375-2c45-8da7-36a1c17bf0ab, 'name': SearchDatastore_Task, 'duration_secs': 0.020675} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.326297] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1129.326564] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1129.326802] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.326965] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.327384] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1129.335802] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f4f14201-946c-48a9-9dd0-1c457b26b9d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.341792] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75879184-5815-410f-a306-d1c8b30f0d55 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.351660] env[65121]: DEBUG oslo_vmware.api [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107320, 'name': PowerOnVM_Task, 'duration_secs': 1.068619} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.354530] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1129.354681] env[65121]: INFO nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Took 12.94 seconds to spawn the instance on the hypervisor. [ 1129.354895] env[65121]: DEBUG nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1129.355656] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2f2fcc-7b4b-471c-ab08-be7100aa902d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.359269] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b256b3f2-a550-4152-a81e-4da3af32ac5a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.364505] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1129.364703] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1129.365956] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1465ce7d-3f49-4aad-8a80-98cc56c3f33f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.383092] env[65121]: DEBUG nova.compute.provider_tree [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.386564] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1129.386564] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d1960c-97d3-a026-c308-30206d18edd8" [ 1129.386564] env[65121]: _type = "Task" [ 1129.386564] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.397207] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d1960c-97d3-a026-c308-30206d18edd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.412940] env[65121]: WARNING openstack [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.413464] env[65121]: WARNING openstack [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.432683] env[65121]: DEBUG nova.network.neutron [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance_info_cache with network_info: [{"id": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "address": "fa:16:3e:14:67:48", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3bafe6b-71", "ovs_interfaceid": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1129.549562] env[65121]: WARNING neutronclient.v2_0.client [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.550242] env[65121]: WARNING openstack [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1129.550580] env[65121]: WARNING openstack [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1129.574191] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquiring lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.574756] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquired lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.574756] env[65121]: DEBUG nova.network.neutron [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1129.706206] env[65121]: DEBUG nova.network.neutron [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updated VIF entry in instance network info cache for port 99be1255-3e71-41e6-936a-acbf44bf9de9. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1129.706662] env[65121]: DEBUG nova.network.neutron [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1129.787060] env[65121]: INFO nova.compute.manager [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Took 23.64 seconds to build instance. [ 1129.886837] env[65121]: DEBUG nova.scheduler.client.report [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1129.901046] env[65121]: INFO nova.compute.manager [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Took 27.13 seconds to build instance. [ 1129.909054] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d1960c-97d3-a026-c308-30206d18edd8, 'name': SearchDatastore_Task, 'duration_secs': 0.011458} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.910490] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e466f3bd-cfbd-475e-8694-2289c5e8d318 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.916958] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1129.916958] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52528c96-fbfc-1ebf-9173-6a807f51239f" [ 1129.916958] env[65121]: _type = "Task" [ 1129.916958] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.925890] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52528c96-fbfc-1ebf-9173-6a807f51239f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.936020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1130.079657] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.080050] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.166789] env[65121]: DEBUG nova.network.neutron [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1130.210100] env[65121]: DEBUG oslo_concurrency.lockutils [req-3521c434-5c5e-465b-af64-0c136d470201 req-d1d1e191-cfdf-41ac-bb55-94b50a2ed65c service nova] Releasing lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1130.289846] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aea47c60-ce90-428f-934b-c94c194bee83 tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.163s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.331919] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "714930d8-3ec9-4aa8-831a-a9f28c365545" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.332420] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.332420] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "714930d8-3ec9-4aa8-831a-a9f28c365545-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.332572] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.332730] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.335640] env[65121]: DEBUG nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1130.336499] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1130.336581] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1130.336773] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1130.336998] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1130.337194] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1130.337415] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1130.337696] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1130.337893] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1130.338298] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1130.338513] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1130.338734] env[65121]: DEBUG nova.virt.hardware [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1130.339575] env[65121]: INFO nova.compute.manager [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Terminating instance [ 1130.341950] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff056843-04c0-4f89-b4cb-243e55ed0da4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.354537] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2168a7-6826-459f-ae77-126ef21391b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.395234] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.965s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.398259] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.941s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.398561] env[65121]: DEBUG nova.objects.instance [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'resources' on Instance uuid e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.404748] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a687368-5a82-4d65-82b1-f7db32a3a9f5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.650s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.431634] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52528c96-fbfc-1ebf-9173-6a807f51239f, 'name': SearchDatastore_Task, 'duration_secs': 0.021321} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.431766] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1130.432858] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1130.432858] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-41098412-949b-4448-83f0-6d95fcc6a496 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.435974] env[65121]: INFO nova.scheduler.client.report [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted allocations for instance fc3109cb-72bb-4695-82c3-a323e619919d [ 1130.446408] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1130.446408] env[65121]: value = "task-5107324" [ 1130.446408] env[65121]: _type = "Task" [ 1130.446408] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.460190] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107324, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.466178] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bbc591-412e-4823-bfbe-d5b567acb450 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.493192] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5c14da-c349-48c0-b905-02cf2e9056aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.502512] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance '9f9118f1-c708-469f-839e-b3c4ef52f38f' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1130.624324] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.624927] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.787809] env[65121]: WARNING neutronclient.v2_0.client [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.787809] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.787809] env[65121]: WARNING openstack [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.851028] env[65121]: DEBUG nova.compute.manager [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1130.851028] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1130.851028] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9597ec2-8448-4fe1-a4b1-7d224ae1726b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.865426] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1130.866644] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3103aba-0c34-4613-aae2-4c569ce0af58 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.880465] env[65121]: DEBUG oslo_vmware.api [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1130.880465] env[65121]: value = "task-5107325" [ 1130.880465] env[65121]: _type = "Task" [ 1130.880465] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.896513] env[65121]: DEBUG oslo_vmware.api [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.945335] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9d6d0462-2b63-45f9-a30c-aa701df4f82e tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "fc3109cb-72bb-4695-82c3-a323e619919d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.504s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.966185] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107324, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.012518] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.012518] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3876e448-cd5e-4210-b89f-ee8d5e724c4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.025251] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1131.025251] env[65121]: value = "task-5107326" [ 1131.025251] env[65121]: _type = "Task" [ 1131.025251] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.037171] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107326, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.221391] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b23a16-994b-43e8-9e76-fe462568eaf2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.230340] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895a74c7-df29-4a8f-831c-ac1de0ce45f3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.269815] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb3cc2f-d925-4bc9-8ff3-4c7343b84cfe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.281098] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a5a242-abf3-4504-bc81-1634876bc8b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.305514] env[65121]: DEBUG nova.compute.provider_tree [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.397898] env[65121]: DEBUG oslo_vmware.api [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107325, 'name': PowerOffVM_Task, 'duration_secs': 0.226382} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.398341] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1131.398550] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1131.398867] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b31b6207-991e-41b4-9e1a-4391bf337f30 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.462636] env[65121]: DEBUG nova.network.neutron [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updating instance_info_cache with network_info: [{"id": "1d9c1075-163b-42c1-b44d-49a4699c9930", "address": "fa:16:3e:96:1f:35", "network": {"id": "fcc1a7c1-12f1-4ecd-a016-509437a5bcf4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1202570986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "529fe6e1d6754df5b4682544b4d3aaa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d9c1075-16", "ovs_interfaceid": "1d9c1075-163b-42c1-b44d-49a4699c9930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1131.470426] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107324, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693736} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.473239] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1131.473239] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1131.473941] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f42960ab-0acc-4dfd-83e7-ecfb4db52d76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.477696] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1131.477696] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1131.477696] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Deleting the datastore file [datastore1] 714930d8-3ec9-4aa8-831a-a9f28c365545 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.477696] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d931062-536c-4a10-b089-0255f56509d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.487188] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1131.487188] env[65121]: value = "task-5107328" [ 1131.487188] env[65121]: _type = "Task" [ 1131.487188] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.487645] env[65121]: DEBUG oslo_vmware.api [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for the task: (returnval){ [ 1131.487645] env[65121]: value = "task-5107329" [ 1131.487645] env[65121]: _type = "Task" [ 1131.487645] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.507479] env[65121]: DEBUG oslo_vmware.api [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.507796] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107328, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.537824] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107326, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.755260] env[65121]: DEBUG nova.compute.manager [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Received event network-changed-1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1131.755866] env[65121]: DEBUG nova.compute.manager [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Refreshing instance network info cache due to event network-changed-1d9c1075-163b-42c1-b44d-49a4699c9930. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1131.757246] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Acquiring lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.811616] env[65121]: DEBUG nova.scheduler.client.report [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1131.972534] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Releasing lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.972890] env[65121]: DEBUG nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Instance network_info: |[{"id": "1d9c1075-163b-42c1-b44d-49a4699c9930", "address": "fa:16:3e:96:1f:35", "network": {"id": "fcc1a7c1-12f1-4ecd-a016-509437a5bcf4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1202570986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "529fe6e1d6754df5b4682544b4d3aaa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d9c1075-16", "ovs_interfaceid": "1d9c1075-163b-42c1-b44d-49a4699c9930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1131.976395] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Acquired lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.976606] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Refreshing network info cache for port 1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1131.979290] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:1f:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3093647a-bab7-4562-ada0-428725e8c0fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d9c1075-163b-42c1-b44d-49a4699c9930', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.988376] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Creating folder: Project (529fe6e1d6754df5b4682544b4d3aaa1). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1131.988376] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65fd1a7d-e552-46e4-9654-b11f5b8be6ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.006922] env[65121]: DEBUG oslo_vmware.api [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Task: {'id': task-5107329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317954} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.009902] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107328, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081395} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.009902] env[65121]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1132.009902] env[65121]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65121) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1132.009902] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.009902] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1132.009902] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1132.009902] env[65121]: INFO nova.compute.manager [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1132.009902] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1132.011311] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1132.011311] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Folder already exists: Project (529fe6e1d6754df5b4682544b4d3aaa1). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1132.011311] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Creating folder: Instances. Parent ref: group-v993549. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1132.011311] env[65121]: DEBUG nova.compute.manager [-] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1132.011311] env[65121]: DEBUG nova.network.neutron [-] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1132.011311] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.011840] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.012110] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.019606] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be616ce1-cad7-4d85-b5d7-7cbadbf37b1a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.022500] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a47ae28-5d46-44bf-af69-a80f55ffe3fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.050888] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1132.055935] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ea15514-d91e-42e2-969e-083bb170aebe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.071474] env[65121]: DEBUG oslo_vmware.api [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107326, 'name': PowerOnVM_Task, 'duration_secs': 0.819058} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.071742] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Created folder: Instances in parent group-v993549. [ 1132.071954] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1132.073024] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.074846] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1132.075043] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8eac6390-ede5-4687-8321-c9dd7963c50e tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance '9f9118f1-c708-469f-839e-b3c4ef52f38f' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.078763] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1132.080270] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4f6186c-1b1e-48ef-baf0-c2e930ecfd12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.098277] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1132.098277] env[65121]: value = "task-5107332" [ 1132.098277] env[65121]: _type = "Task" [ 1132.098277] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.104784] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1132.104784] env[65121]: value = "task-5107333" [ 1132.104784] env[65121]: _type = "Task" [ 1132.104784] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.108785] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107332, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.118553] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107333, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.318715] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.920s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.321993] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.792s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.347909] env[65121]: INFO nova.scheduler.client.report [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted allocations for instance e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2 [ 1132.479298] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.480032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.488248] env[65121]: WARNING neutronclient.v2_0.client [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.489016] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.489434] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.621125] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107332, 'name': ReconfigVM_Task, 'duration_secs': 0.312548} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.621817] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfigured VM instance instance-0000006c to attach disk [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.622570] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c87b82cf-fe11-4394-a4f5-97fe18fd2804 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.630534] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107333, 'name': CreateVM_Task, 'duration_secs': 0.392564} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.630534] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1132.630534] env[65121]: WARNING neutronclient.v2_0.client [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.630850] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'attachment_id': '219e8d7a-c5fa-4b8f-a91a-e971ee2b447d', 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993557', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'name': 'volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4f86672c-425d-4641-a68c-2d6f5324e382', 'attached_at': '', 'detached_at': '', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'serial': '20d5c722-822e-46f0-9b9b-7f7d92bd114a'}, 'mount_device': '/dev/sda', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65121) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1132.630930] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Root volume attach. Driver type: vmdk {{(pid=65121) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1132.632021] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72523a16-976c-427b-b854-f1c53600bec1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.638563] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1132.638563] env[65121]: value = "task-5107334" [ 1132.638563] env[65121]: _type = "Task" [ 1132.638563] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.651441] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794d383c-0a13-44e8-a1fe-60a8103307a7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.658501] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107334, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.664405] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694ad984-a8aa-4e9a-a90b-1dba81062d5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.672689] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-6344682a-6b49-4e9b-aad2-983b656aff8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.681426] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1132.681426] env[65121]: value = "task-5107335" [ 1132.681426] env[65121]: _type = "Task" [ 1132.681426] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.686649] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.687046] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.700207] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107335, 'name': RelocateVM_Task} progress is 7%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.780928] env[65121]: DEBUG nova.compute.manager [req-208d0403-c96f-4200-9e15-85413b41b2e4 req-99cec924-0096-4f25-b19f-54445e6c98b7 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Received event network-vif-deleted-a08a0ecc-509c-49a0-923d-4f9caf3e193e {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1132.781279] env[65121]: INFO nova.compute.manager [req-208d0403-c96f-4200-9e15-85413b41b2e4 req-99cec924-0096-4f25-b19f-54445e6c98b7 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Neutron deleted interface a08a0ecc-509c-49a0-923d-4f9caf3e193e; detaching it from the instance and deleting it from the info cache [ 1132.781563] env[65121]: DEBUG nova.network.neutron [req-208d0403-c96f-4200-9e15-85413b41b2e4 req-99cec924-0096-4f25-b19f-54445e6c98b7 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1132.828591] env[65121]: WARNING neutronclient.v2_0.client [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.829325] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.829760] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.863023] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e828f277-2a44-4c80-a976-ba134d9e91a7 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.827s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.904850] env[65121]: DEBUG nova.network.neutron [-] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1132.930204] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updated VIF entry in instance network info cache for port 1d9c1075-163b-42c1-b44d-49a4699c9930. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1132.930638] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updating instance_info_cache with network_info: [{"id": "1d9c1075-163b-42c1-b44d-49a4699c9930", "address": "fa:16:3e:96:1f:35", "network": {"id": "fcc1a7c1-12f1-4ecd-a016-509437a5bcf4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1202570986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "529fe6e1d6754df5b4682544b4d3aaa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d9c1075-16", "ovs_interfaceid": "1d9c1075-163b-42c1-b44d-49a4699c9930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1132.983873] env[65121]: DEBUG nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1133.093216] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.093463] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.150924] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107334, 'name': Rename_Task, 'duration_secs': 0.160311} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.151208] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1133.151462] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-487c0b63-ff20-467b-8ef9-ee6cdba8e2be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.160366] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1133.160366] env[65121]: value = "task-5107336" [ 1133.160366] env[65121]: _type = "Task" [ 1133.160366] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.172456] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.193082] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107335, 'name': RelocateVM_Task, 'duration_secs': 0.029837} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.193254] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1133.193453] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993557', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'name': 'volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4f86672c-425d-4641-a68c-2d6f5324e382', 'attached_at': '', 'detached_at': '', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'serial': '20d5c722-822e-46f0-9b9b-7f7d92bd114a'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1133.194307] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa21ca1-ee29-4c0e-966d-08aaca7e0085 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.213493] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f46e715-c0da-4f3b-8c96-ffc70a0db2a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.239825] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a/volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.240499] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56e26f3a-947e-4f3d-b182-6c7a82c671fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.261953] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1133.261953] env[65121]: value = "task-5107337" [ 1133.261953] env[65121]: _type = "Task" [ 1133.261953] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.273106] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107337, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.284945] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6149ea3a-1d6f-490e-b93d-0a49adc82e1a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.296162] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6db7e1b-e12b-4d69-bc28-ad6f1362b152 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.336423] env[65121]: DEBUG nova.compute.manager [req-208d0403-c96f-4200-9e15-85413b41b2e4 req-99cec924-0096-4f25-b19f-54445e6c98b7 service nova] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Detach interface failed, port_id=a08a0ecc-509c-49a0-923d-4f9caf3e193e, reason: Instance 714930d8-3ec9-4aa8-831a-a9f28c365545 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1133.346899] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Applying migration context for instance 9f9118f1-c708-469f-839e-b3c4ef52f38f as it has an incoming, in-progress migration f186477a-a599-4c02-99d7-f463c3b63131. Migration status is finished {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1133.348648] env[65121]: INFO nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating resource usage from migration c73f3217-4588-4e9d-b62a-5adf5dc65cf2 [ 1133.348893] env[65121]: INFO nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating resource usage from migration f186477a-a599-4c02-99d7-f463c3b63131 [ 1133.368395] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d348e1-7ae8-4d95-a6b6-e6834291189d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.376044] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a28074-cf57-4caf-9a9e-6579dfc745f1 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Suspending the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1133.376928] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f84d7aa1-a90b-4c95-9673-fb1a2af7187c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377086] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4a72fedd-b114-468e-8f34-0caec6ce73fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377241] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 6fe86deb-0903-4769-a05c-b7d0acec0103 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377971] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 59ceb0fd-1fb5-4c90-963d-fe76b9740d29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377971] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377971] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377971] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377971] env[65121]: WARNING nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 0497d253-887b-47a9-be8d-cc4592e046ba is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1133.377971] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 8fce40bb-69c1-4fbb-9b0b-c7aece54d179 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.377971] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Migration f186477a-a599-4c02-99d7-f463c3b63131 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1133.378334] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 9f9118f1-c708-469f-839e-b3c4ef52f38f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.378334] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 714930d8-3ec9-4aa8-831a-a9f28c365545 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.378334] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f58bf796-69e7-4a61-9cea-78dff8d4fab5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.378453] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4f86672c-425d-4641-a68c-2d6f5324e382 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.378566] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Migration c73f3217-4588-4e9d-b62a-5adf5dc65cf2 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1133.378637] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 84ae6061-a962-4a74-8661-9718cc4c5346 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1133.379994] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-1cb0de06-d3e4-4051-8e93-5cad6a9b0e56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.389470] env[65121]: DEBUG oslo_vmware.api [None req-e3a28074-cf57-4caf-9a9e-6579dfc745f1 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1133.389470] env[65121]: value = "task-5107338" [ 1133.389470] env[65121]: _type = "Task" [ 1133.389470] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.401736] env[65121]: DEBUG oslo_vmware.api [None req-e3a28074-cf57-4caf-9a9e-6579dfc745f1 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107338, 'name': SuspendVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.407391] env[65121]: INFO nova.compute.manager [-] [instance: 714930d8-3ec9-4aa8-831a-a9f28c365545] Took 1.40 seconds to deallocate network for instance. [ 1133.435078] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Releasing lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.435401] env[65121]: DEBUG nova.compute.manager [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1133.435566] env[65121]: DEBUG nova.compute.manager [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing instance network info cache due to event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1133.435975] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.436082] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.436281] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1133.512586] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.603965] env[65121]: INFO nova.compute.manager [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Detaching volume 7bcf71f4-0c34-441a-9533-931791d14ff6 [ 1133.652741] env[65121]: INFO nova.virt.block_device [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Attempting to driver detach volume 7bcf71f4-0c34-441a-9533-931791d14ff6 from mountpoint /dev/sdb [ 1133.652982] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1133.653181] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993517', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'name': 'volume-7bcf71f4-0c34-441a-9533-931791d14ff6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59ceb0fd-1fb5-4c90-963d-fe76b9740d29', 'attached_at': '', 'detached_at': '', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'serial': '7bcf71f4-0c34-441a-9533-931791d14ff6'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1133.654264] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b703899-5794-4352-85d8-c568879cec81 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.688286] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b7775c-6809-47b9-b676-855eadc00f74 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.695246] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107336, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.701900] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d34eb8a-add6-49be-a2a2-f7aeb7ddd148 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.727252] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa807c84-ee74-42c3-8c6f-c5d02aee595d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.747171] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The volume has not been displaced from its original location: [datastore1] volume-7bcf71f4-0c34-441a-9533-931791d14ff6/volume-7bcf71f4-0c34-441a-9533-931791d14ff6.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1133.752699] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.753154] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d6abc41-c85d-40e6-a1a7-473b5effdf5b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.776928] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107337, 'name': ReconfigVM_Task, 'duration_secs': 0.34236} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.778544] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a/volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1133.783580] env[65121]: DEBUG oslo_vmware.api [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1133.783580] env[65121]: value = "task-5107339" [ 1133.783580] env[65121]: _type = "Task" [ 1133.783580] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.783979] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9335eea0-3403-4a4b-8fbe-de26266a29ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.804197] env[65121]: DEBUG nova.compute.manager [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1133.804357] env[65121]: DEBUG nova.compute.manager [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing instance network info cache due to event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1133.804616] env[65121]: DEBUG oslo_concurrency.lockutils [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.804752] env[65121]: DEBUG oslo_concurrency.lockutils [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.804902] env[65121]: DEBUG nova.network.neutron [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing network info cache for port ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1133.816659] env[65121]: DEBUG oslo_vmware.api [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107339, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.818565] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1133.818565] env[65121]: value = "task-5107340" [ 1133.818565] env[65121]: _type = "Task" [ 1133.818565] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.831667] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107340, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.883625] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 999ab257-c116-4a73-9b87-b52aeb9a3bb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1133.883625] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1133.883912] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=100GB used_disk=13GB total_vcpus=48 used_vcpus=14 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '13', 'num_vm_active': '11', 'num_task_None': '8', 'num_os_type_None': '13', 'num_proj_89110618b6ef4ce68b2ca53e7e168139': '1', 'io_workload': '3', 'num_proj_00383aa3355e438cb703c2b86c7917f2': '2', 'num_proj_48809b7bf36c4315b9ac315d6b9ca0c8': '1', 'num_proj_1ce7f6698e214d73ae43427601058af8': '1', 'num_proj_f056059180af48bba4587006efb0b211': '1', 'num_task_resize_prep': '1', 'num_proj_4467da3ed41245ddbc93fc865a8b7bdd': '1', 'num_proj_35312a302644426f98f127e89a067e75': '2', 'num_task_resize_finish': '1', 'num_proj_e491b88dcaad42dcb345b3217bc6f941': '1', 'num_task_deleting': '1', 'num_proj_20d977ed0564461894b3774afbf96a3a': '1', 'num_vm_building': '2', 'num_task_spawning': '2', 'num_proj_72237a0d762645588c41231b0a34a796': '1', 'num_proj_529fe6e1d6754df5b4682544b4d3aaa1': '1'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1133.905763] env[65121]: DEBUG oslo_vmware.api [None req-e3a28074-cf57-4caf-9a9e-6579dfc745f1 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107338, 'name': SuspendVM_Task} progress is 62%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.916731] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.938704] env[65121]: WARNING neutronclient.v2_0.client [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.939533] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.939907] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.113748] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.114151] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.129033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.129033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.174309] env[65121]: DEBUG oslo_vmware.api [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107336, 'name': PowerOnVM_Task, 'duration_secs': 0.580141} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.174549] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1134.174744] env[65121]: INFO nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Took 8.06 seconds to spawn the instance on the hypervisor. [ 1134.174910] env[65121]: DEBUG nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1134.175833] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd3a67f-799d-4ea9-9df8-f54f6fdd0554 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.191948] env[65121]: WARNING neutronclient.v2_0.client [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.192653] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.193030] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.218395] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723f4d38-7f90-449a-8cd8-f71bdad3a2cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.228096] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126719f0-1812-4ade-a1d6-c66047c55ae8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.268193] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a9ab3f-9e09-4958-a40c-27af2cb17644 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.278209] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9eaa35-8711-41c0-9dbb-02f1d35a5180 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.299096] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.311582] env[65121]: WARNING neutronclient.v2_0.client [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.312402] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.312856] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.320825] env[65121]: DEBUG oslo_vmware.api [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107339, 'name': ReconfigVM_Task, 'duration_secs': 0.284703} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.323142] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updated VIF entry in instance network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1134.323515] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.326040] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1134.341697] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed7d629f-9fbc-4648-89b7-6bd18a0e3bcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.375316] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107340, 'name': ReconfigVM_Task, 'duration_secs': 0.17166} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.380441] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993557', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'name': 'volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4f86672c-425d-4641-a68c-2d6f5324e382', 'attached_at': '', 'detached_at': '', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'serial': '20d5c722-822e-46f0-9b9b-7f7d92bd114a'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1134.381381] env[65121]: DEBUG oslo_vmware.api [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1134.381381] env[65121]: value = "task-5107341" [ 1134.381381] env[65121]: _type = "Task" [ 1134.381381] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.382327] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45494be3-6ebd-4f03-b51e-ca0c8f446f49 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.396285] env[65121]: DEBUG oslo_vmware.api [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107341, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.397455] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1134.397455] env[65121]: value = "task-5107342" [ 1134.397455] env[65121]: _type = "Task" [ 1134.397455] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.414197] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107342, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.414529] env[65121]: DEBUG oslo_vmware.api [None req-e3a28074-cf57-4caf-9a9e-6579dfc745f1 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107338, 'name': SuspendVM_Task, 'duration_secs': 0.929149} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.414694] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a28074-cf57-4caf-9a9e-6579dfc745f1 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Suspended the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1134.414863] env[65121]: DEBUG nova.compute.manager [None req-e3a28074-cf57-4caf-9a9e-6579dfc745f1 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1134.415808] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5358a6-5ead-430d-9451-8afbad4b4b63 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.468889] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.469339] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.554028] env[65121]: WARNING neutronclient.v2_0.client [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.554028] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.554028] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.629086] env[65121]: DEBUG nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1134.646710] env[65121]: DEBUG nova.network.neutron [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updated VIF entry in instance network info cache for port ff1a1ddf-0705-4111-80a1-faa79fe7c947. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1134.647165] env[65121]: DEBUG nova.network.neutron [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.698662] env[65121]: INFO nova.compute.manager [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Took 23.66 seconds to build instance. [ 1134.805211] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1134.836431] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1134.836737] env[65121]: DEBUG nova.compute.manager [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1134.836900] env[65121]: DEBUG nova.compute.manager [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing instance network info cache due to event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1134.837095] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.869998] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.870545] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.870765] env[65121]: DEBUG nova.compute.manager [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Going to confirm migration 6 {{(pid=65121) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1134.897024] env[65121]: DEBUG oslo_vmware.api [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107341, 'name': ReconfigVM_Task, 'duration_secs': 0.189837} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.897351] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993517', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'name': 'volume-7bcf71f4-0c34-441a-9533-931791d14ff6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59ceb0fd-1fb5-4c90-963d-fe76b9740d29', 'attached_at': '', 'detached_at': '', 'volume_id': '7bcf71f4-0c34-441a-9533-931791d14ff6', 'serial': '7bcf71f4-0c34-441a-9533-931791d14ff6'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1134.910595] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107342, 'name': Rename_Task, 'duration_secs': 0.187907} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.910854] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1134.911117] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2573ae31-826c-464c-b849-3a6cff9655ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.918945] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1134.918945] env[65121]: value = "task-5107343" [ 1134.918945] env[65121]: _type = "Task" [ 1134.918945] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.930208] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107343, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.150924] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.151631] env[65121]: DEBUG oslo_concurrency.lockutils [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.152020] env[65121]: DEBUG nova.compute.manager [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1135.152324] env[65121]: DEBUG nova.compute.manager [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing instance network info cache due to event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1135.152665] env[65121]: DEBUG oslo_concurrency.lockutils [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.152917] env[65121]: DEBUG oslo_concurrency.lockutils [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.153200] env[65121]: DEBUG nova.network.neutron [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1135.156789] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.157377] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing network info cache for port ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1135.201051] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b94dcfba-ed99-4a9e-8d75-5151595635b9 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.170s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.302876] env[65121]: DEBUG nova.compute.manager [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Received event network-changed-99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1135.304253] env[65121]: DEBUG nova.compute.manager [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Refreshing instance network info cache due to event network-changed-99be1255-3e71-41e6-936a-acbf44bf9de9. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1135.304253] env[65121]: DEBUG oslo_concurrency.lockutils [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Acquiring lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.304253] env[65121]: DEBUG oslo_concurrency.lockutils [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Acquired lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.304253] env[65121]: DEBUG nova.network.neutron [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Refreshing network info cache for port 99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1135.310848] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1135.311026] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.989s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.311283] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.912s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.311453] env[65121]: DEBUG nova.objects.instance [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65121) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1135.313999] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.314165] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Cleaning up deleted instances {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 1135.378906] env[65121]: WARNING neutronclient.v2_0.client [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.434438] env[65121]: DEBUG oslo_vmware.api [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107343, 'name': PowerOnVM_Task, 'duration_secs': 0.479863} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.434759] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1135.434995] env[65121]: INFO nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Took 5.10 seconds to spawn the instance on the hypervisor. [ 1135.435224] env[65121]: DEBUG nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1135.436127] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565152a7-69b4-4acd-9fe1-9991b27cd5a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.439820] env[65121]: WARNING neutronclient.v2_0.client [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.440292] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.440292] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.440430] env[65121]: DEBUG nova.network.neutron [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1135.440582] env[65121]: DEBUG nova.objects.instance [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lazy-loading 'info_cache' on Instance uuid 9f9118f1-c708-469f-839e-b3c4ef52f38f {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.452788] env[65121]: DEBUG nova.objects.instance [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'flavor' on Instance uuid 59ceb0fd-1fb5-4c90-963d-fe76b9740d29 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.659970] env[65121]: WARNING neutronclient.v2_0.client [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.660877] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.661291] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.669457] env[65121]: WARNING neutronclient.v2_0.client [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.670072] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.670466] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.762914] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.762914] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.773977] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.774375] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.806648] env[65121]: WARNING neutronclient.v2_0.client [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.807293] env[65121]: WARNING openstack [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.807639] env[65121]: WARNING openstack [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.833622] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] There are 53 instances to clean {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 1135.834386] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: e445dcc1-f0c1-48b2-bce9-ee0d657dd2b2] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1135.838270] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76ad2326-521c-48e3-b069-2d80af652768 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.527s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.843780] env[65121]: WARNING neutronclient.v2_0.client [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.843780] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.843780] env[65121]: WARNING openstack [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.855132] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.487s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.855358] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.857236] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 12.031s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.864094] env[65121]: WARNING neutronclient.v2_0.client [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.865601] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.865601] env[65121]: WARNING openstack [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.887284] env[65121]: INFO nova.scheduler.client.report [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted allocations for instance 0497d253-887b-47a9-be8d-cc4592e046ba [ 1135.906170] env[65121]: INFO nova.compute.manager [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Resuming [ 1135.906825] env[65121]: DEBUG nova.objects.instance [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lazy-loading 'flavor' on Instance uuid 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.964098] env[65121]: INFO nova.compute.manager [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Took 22.98 seconds to build instance. [ 1135.969827] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updated VIF entry in instance network info cache for port ff1a1ddf-0705-4111-80a1-faa79fe7c947. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1135.970278] env[65121]: DEBUG nova.network.neutron [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1135.972718] env[65121]: WARNING openstack [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.973081] env[65121]: WARNING openstack [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.013228] env[65121]: DEBUG nova.network.neutron [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updated VIF entry in instance network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1136.013623] env[65121]: DEBUG nova.network.neutron [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.035289] env[65121]: WARNING neutronclient.v2_0.client [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.036022] env[65121]: WARNING openstack [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.036449] env[65121]: WARNING openstack [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.128873] env[65121]: DEBUG nova.network.neutron [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updated VIF entry in instance network info cache for port 99be1255-3e71-41e6-936a-acbf44bf9de9. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1136.129285] env[65121]: DEBUG nova.network.neutron [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.130438] env[65121]: WARNING oslo_messaging._drivers.amqpdriver [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 1136.339808] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 39ca7a89-c262-4169-91b1-92ea45744518] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1136.361742] env[65121]: INFO nova.compute.claims [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1136.395661] env[65121]: DEBUG oslo_concurrency.lockutils [None req-2ea90f48-313e-4369-a20d-c7aa37f1fff0 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "0497d253-887b-47a9-be8d-cc4592e046ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.135s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.448528] env[65121]: WARNING neutronclient.v2_0.client [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.449279] env[65121]: WARNING openstack [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.449630] env[65121]: WARNING openstack [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.466029] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c27b00a6-0b09-49c7-a530-0c25ccd198ed tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.372s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.467919] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0ced5e0b-5ec8-47b8-a789-5b2ab6c74b1b tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "4f86672c-425d-4641-a68c-2d6f5324e382" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.489s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.482187] env[65121]: DEBUG oslo_concurrency.lockutils [req-b231fa21-7e5f-4c6e-b688-6eb90e0695de req-598c9118-8796-4cc7-b1d3-9c52d6285711 service nova] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.515888] env[65121]: DEBUG oslo_concurrency.lockutils [req-75ca79de-df9e-4858-831a-0cd1637782b1 req-8aed4d2e-9210-48ed-81ab-d4a9e6d70137 service nova] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.567696] env[65121]: WARNING openstack [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.567779] env[65121]: WARNING openstack [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.607153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.607483] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.607692] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.607852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.608116] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.610381] env[65121]: INFO nova.compute.manager [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Terminating instance [ 1136.633945] env[65121]: DEBUG oslo_concurrency.lockutils [req-b045f0f0-9d49-4c61-8f1b-a938dd652e33 req-2dd4960a-90bf-45ba-94c7-1418f10bf344 service nova] Releasing lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.684022] env[65121]: WARNING neutronclient.v2_0.client [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.684022] env[65121]: WARNING openstack [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.684022] env[65121]: WARNING openstack [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.844132] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 347bbb23-da89-4e6f-837c-e88bf7c57a3a] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1136.863626] env[65121]: DEBUG nova.network.neutron [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance_info_cache with network_info: [{"id": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "address": "fa:16:3e:14:67:48", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3bafe6b-71", "ovs_interfaceid": "b3bafe6b-71bc-45d0-bf37-29da60b15de0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.869487] env[65121]: INFO nova.compute.resource_tracker [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating resource usage from migration c73f3217-4588-4e9d-b62a-5adf5dc65cf2 [ 1137.114677] env[65121]: DEBUG nova.compute.manager [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1137.114889] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.115847] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a95973-44fe-440a-a9db-ce3893963c55 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.119981] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ea59b3-a280-4ac1-8e12-3650485ebe13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.129887] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ab5cd7-b8e2-4cdc-a091-b5b62d1ed0e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.132919] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.133157] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0146fb3-ca38-4be8-ac09-571962538fd3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.164576] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7742a6b-3bfc-4d98-a560-f5c171e31c1e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.167557] env[65121]: DEBUG oslo_vmware.api [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1137.167557] env[65121]: value = "task-5107344" [ 1137.167557] env[65121]: _type = "Task" [ 1137.167557] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.175277] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddfdb7a-16ad-476c-9d59-ba58c4411207 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.182590] env[65121]: DEBUG oslo_vmware.api [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107344, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.195632] env[65121]: DEBUG nova.compute.provider_tree [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.347232] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 97fa5791-28cb-4e77-bad7-f2cbce8bd06d] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1137.366341] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-9f9118f1-c708-469f-839e-b3c4ef52f38f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.366698] env[65121]: DEBUG nova.objects.instance [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lazy-loading 'migration_context' on Instance uuid 9f9118f1-c708-469f-839e-b3c4ef52f38f {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.388703] env[65121]: DEBUG nova.compute.manager [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Received event network-changed-1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1137.388976] env[65121]: DEBUG nova.compute.manager [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Refreshing instance network info cache due to event network-changed-1d9c1075-163b-42c1-b44d-49a4699c9930. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1137.389559] env[65121]: DEBUG oslo_concurrency.lockutils [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Acquiring lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.389559] env[65121]: DEBUG oslo_concurrency.lockutils [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Acquired lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1137.389559] env[65121]: DEBUG nova.network.neutron [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Refreshing network info cache for port 1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1137.422746] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.423154] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquired lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1137.423453] env[65121]: DEBUG nova.network.neutron [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1137.678951] env[65121]: DEBUG oslo_vmware.api [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107344, 'name': PowerOffVM_Task, 'duration_secs': 0.410493} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.679273] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1137.679439] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1137.679693] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6252e18d-fd6b-4dff-b7cb-c53eb5df96ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.698722] env[65121]: DEBUG nova.scheduler.client.report [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1137.753470] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1137.753747] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1137.754019] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleting the datastore file [datastore2] 59ceb0fd-1fb5-4c90-963d-fe76b9740d29 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.754654] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40fb70aa-71f5-4805-92ca-b7fea40c8855 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.763481] env[65121]: DEBUG oslo_vmware.api [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1137.763481] env[65121]: value = "task-5107346" [ 1137.763481] env[65121]: _type = "Task" [ 1137.763481] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.772644] env[65121]: DEBUG oslo_vmware.api [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.850401] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 03a13029-55b1-4757-80a6-9293471434dc] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1137.870243] env[65121]: DEBUG nova.objects.base [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Object Instance<9f9118f1-c708-469f-839e-b3c4ef52f38f> lazy-loaded attributes: info_cache,migration_context {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1137.871252] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553e2568-1c56-4e83-b15a-1ac8a786d1ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.893178] env[65121]: WARNING neutronclient.v2_0.client [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.893900] env[65121]: WARNING openstack [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.894294] env[65121]: WARNING openstack [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.903301] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-546e930f-93e6-4973-897c-c3864f6dfe3b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.911490] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.911812] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1137.914992] env[65121]: DEBUG oslo_vmware.api [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1137.914992] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e0a6da-b055-8f62-04ec-eaffb31bc4bc" [ 1137.914992] env[65121]: _type = "Task" [ 1137.914992] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.926487] env[65121]: DEBUG oslo_vmware.api [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e0a6da-b055-8f62-04ec-eaffb31bc4bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009135} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.926977] env[65121]: WARNING neutronclient.v2_0.client [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.927604] env[65121]: WARNING openstack [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.928033] env[65121]: WARNING openstack [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.935397] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.041787] env[65121]: WARNING openstack [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.041950] env[65121]: WARNING openstack [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.058822] env[65121]: WARNING openstack [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.060036] env[65121]: WARNING openstack [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.123539] env[65121]: WARNING neutronclient.v2_0.client [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.124104] env[65121]: WARNING openstack [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.124493] env[65121]: WARNING openstack [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.146170] env[65121]: WARNING neutronclient.v2_0.client [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.146897] env[65121]: WARNING openstack [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.147348] env[65121]: WARNING openstack [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.203717] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.346s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.203944] env[65121]: INFO nova.compute.manager [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Migrating [ 1138.212088] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.700s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.213774] env[65121]: INFO nova.compute.claims [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1138.221109] env[65121]: DEBUG nova.network.neutron [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updated VIF entry in instance network info cache for port 1d9c1075-163b-42c1-b44d-49a4699c9930. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1138.221518] env[65121]: DEBUG nova.network.neutron [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updating instance_info_cache with network_info: [{"id": "1d9c1075-163b-42c1-b44d-49a4699c9930", "address": "fa:16:3e:96:1f:35", "network": {"id": "fcc1a7c1-12f1-4ecd-a016-509437a5bcf4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1202570986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "529fe6e1d6754df5b4682544b4d3aaa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d9c1075-16", "ovs_interfaceid": "1d9c1075-163b-42c1-b44d-49a4699c9930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1138.230138] env[65121]: DEBUG nova.network.neutron [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [{"id": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "address": "fa:16:3e:c9:95:2b", "network": {"id": "808c43d5-1b16-4fbf-ae95-55dfe12079d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2118700129-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48809b7bf36c4315b9ac315d6b9ca0c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04d47dd2-7f", "ovs_interfaceid": "04d47dd2-7f48-48e4-aed4-9d693168b8ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1138.275136] env[65121]: DEBUG oslo_vmware.api [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217103} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.275777] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.275991] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.276346] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.276488] env[65121]: INFO nova.compute.manager [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1138.276729] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1138.276956] env[65121]: DEBUG nova.compute.manager [-] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1138.277111] env[65121]: DEBUG nova.network.neutron [-] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1138.277332] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.277890] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.278257] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.341874] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.355097] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: fc3109cb-72bb-4695-82c3-a323e619919d] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1138.416687] env[65121]: DEBUG nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1138.488900] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.489272] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.489414] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "6fe86deb-0903-4769-a05c-b7d0acec0103-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.489593] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.489757] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.495110] env[65121]: INFO nova.compute.manager [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Terminating instance [ 1138.729493] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.729493] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1138.729493] env[65121]: DEBUG nova.network.neutron [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1138.730023] env[65121]: DEBUG oslo_concurrency.lockutils [req-ecdb8dac-bf03-4115-b65e-af88470d42f7 req-e1b12749-393f-4134-8af8-c87baf11caf7 service nova] Releasing lock "refresh_cache-4f86672c-425d-4641-a68c-2d6f5324e382" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1138.733298] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Releasing lock "refresh_cache-52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1138.734945] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3678db75-6a41-4191-a0d5-35e000c9956c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.743948] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Resuming the VM {{(pid=65121) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1138.744473] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f9a7567-b94a-4a23-93be-6a3beb14b2f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.753744] env[65121]: DEBUG oslo_vmware.api [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1138.753744] env[65121]: value = "task-5107347" [ 1138.753744] env[65121]: _type = "Task" [ 1138.753744] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.763074] env[65121]: DEBUG oslo_vmware.api [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.858489] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 3d6fc9df-6872-4929-8813-bec1db1ebb16] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1138.940625] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.004253] env[65121]: DEBUG nova.compute.manager [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1139.004468] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1139.005364] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed929396-2dd7-44db-802f-1f4bf9c1788c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.013690] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1139.013895] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ce89b17-5c75-4295-b657-a008fa168d41 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.021580] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1139.021580] env[65121]: value = "task-5107348" [ 1139.021580] env[65121]: _type = "Task" [ 1139.021580] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.031807] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107348, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.093901] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.093901] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.094281] env[65121]: DEBUG nova.objects.instance [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'flavor' on Instance uuid 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.234225] env[65121]: WARNING neutronclient.v2_0.client [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.234880] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.235249] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.264452] env[65121]: DEBUG oslo_vmware.api [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107347, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.334965] env[65121]: DEBUG nova.network.neutron [-] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1139.363065] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 11caa171-0314-4e7d-8502-80932b469da8] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1139.367548] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.368011] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.470260] env[65121]: DEBUG nova.compute.manager [req-d291e65e-6771-4d2d-a589-780f8c0f7091 req-c6671402-a55d-4e0f-b16b-02b7c4c84670 service nova] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Received event network-vif-deleted-ad549932-0308-4165-96b1-9b4e419b6d05 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1139.482392] env[65121]: WARNING neutronclient.v2_0.client [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.483850] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.483944] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.536995] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107348, 'name': PowerOffVM_Task, 'duration_secs': 0.415692} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.540176] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1139.540410] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1139.541388] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e7c86b4-9b73-4add-b918-9350ce4f04ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.560380] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5b4159-6af9-4a97-b7b8-14b253792194 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.571219] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab06885-c0f3-4d24-a6d9-ffe60a2e0cdd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.609240] env[65121]: WARNING neutronclient.v2_0.client [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.609954] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.610325] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.620630] env[65121]: DEBUG nova.network.neutron [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1139.622959] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8d56af-1fc0-45cb-914a-823f488d0265 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.627486] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1139.627696] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1139.627869] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Deleting the datastore file [datastore2] 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1139.628654] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa0caffd-a2a5-4b55-af9e-e0a94a214736 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.637480] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3b51ed-02d5-48a5-8db0-5ed7b560ebf1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.643830] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1139.643830] env[65121]: value = "task-5107350" [ 1139.643830] env[65121]: _type = "Task" [ 1139.643830] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.657681] env[65121]: DEBUG nova.compute.provider_tree [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.672670] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.753588] env[65121]: DEBUG nova.objects.instance [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'pci_requests' on Instance uuid 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.768170] env[65121]: DEBUG oslo_vmware.api [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107347, 'name': PowerOnVM_Task, 'duration_secs': 0.888045} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.768499] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Resumed the VM {{(pid=65121) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1139.768627] env[65121]: DEBUG nova.compute.manager [None req-5e93545e-b2b1-4290-89ec-1fc77b475359 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1139.769490] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998ee865-2185-4b8e-a20a-ff16de0d1c91 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.839163] env[65121]: INFO nova.compute.manager [-] [instance: 59ceb0fd-1fb5-4c90-963d-fe76b9740d29] Took 1.56 seconds to deallocate network for instance. [ 1139.883481] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 76e4cddb-b046-4a3e-9a91-d3ee847883cf] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1140.127753] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1140.156404] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.165489] env[65121]: DEBUG nova.scheduler.client.report [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1140.260356] env[65121]: DEBUG nova.objects.base [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Object Instance<3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7> lazy-loaded attributes: flavor,pci_requests {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1140.260599] env[65121]: DEBUG nova.network.neutron [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1140.260951] env[65121]: WARNING neutronclient.v2_0.client [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.261292] env[65121]: WARNING neutronclient.v2_0.client [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.261821] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.262183] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.336145] env[65121]: DEBUG nova.policy [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1140.346549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.386912] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: abd87ca7-5d95-4b22-8d2a-1f37abd18f86] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1140.655038] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.671091] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.671630] env[65121]: DEBUG nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1140.674906] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.758s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.675146] env[65121]: DEBUG nova.objects.instance [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lazy-loading 'resources' on Instance uuid 714930d8-3ec9-4aa8-831a-a9f28c365545 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.891042] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: ec463cad-4c80-4636-bc7a-9ec298a07d96] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1141.156366] env[65121]: DEBUG oslo_vmware.api [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.068824} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.156632] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1141.156805] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1141.156970] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1141.157148] env[65121]: INFO nova.compute.manager [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Took 2.15 seconds to destroy the instance on the hypervisor. [ 1141.157388] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1141.157861] env[65121]: DEBUG nova.compute.manager [-] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1141.157984] env[65121]: DEBUG nova.network.neutron [-] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1141.158323] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.158862] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.159137] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.178695] env[65121]: DEBUG nova.compute.utils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1141.183050] env[65121]: DEBUG nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1141.183265] env[65121]: DEBUG nova.network.neutron [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1141.183588] env[65121]: WARNING neutronclient.v2_0.client [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.183848] env[65121]: WARNING neutronclient.v2_0.client [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.184464] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.184819] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.197666] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.270780] env[65121]: DEBUG nova.policy [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10dd394ea47044e0ad78daacb6aa161a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad1aea30d62c45e193c5a54c429ce7ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1141.394371] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 4af106e2-0a8a-4e93-91e7-2fbbcadc5ad3] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1141.449758] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ac7db9-649b-4519-a6e3-8b7c5b8207f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.458843] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1208c83-0774-4197-9c80-6556f797a31b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.495964] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c9936f-0598-4fb8-a773-217c64c382ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.504852] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beada4db-54e7-4d40-a9b8-a205d154c0c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.519946] env[65121]: DEBUG nova.compute.provider_tree [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.643951] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8e4d8e-6340-4760-926d-18dc4bf9d14e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.667770] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance '84ae6061-a962-4a74-8661-9718cc4c5346' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1141.683491] env[65121]: DEBUG nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1141.881496] env[65121]: DEBUG nova.network.neutron [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Successfully updated port: f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1141.901728] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: e756fdb2-f901-4e23-92e7-f2753fd5f728] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1142.024882] env[65121]: DEBUG nova.scheduler.client.report [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.071605] env[65121]: DEBUG nova.network.neutron [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Successfully created port: 91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1142.147614] env[65121]: DEBUG nova.compute.manager [req-6a811323-6328-414b-a195-8b950ded5b97 req-7a40f0cb-a096-4821-8af9-8ea7f048c89a service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Received event network-vif-deleted-e31cbeae-05bc-416d-9df3-10f09d947ba4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1142.147906] env[65121]: INFO nova.compute.manager [req-6a811323-6328-414b-a195-8b950ded5b97 req-7a40f0cb-a096-4821-8af9-8ea7f048c89a service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Neutron deleted interface e31cbeae-05bc-416d-9df3-10f09d947ba4; detaching it from the instance and deleting it from the info cache [ 1142.148203] env[65121]: DEBUG nova.network.neutron [req-6a811323-6328-414b-a195-8b950ded5b97 req-7a40f0cb-a096-4821-8af9-8ea7f048c89a service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1142.168036] env[65121]: DEBUG nova.compute.manager [req-ddd1ca7f-814a-4749-b25b-577325d23704 req-462a5fb7-12e8-4619-8443-7a6c4f70f7c0 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-vif-plugged-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1142.168275] env[65121]: DEBUG oslo_concurrency.lockutils [req-ddd1ca7f-814a-4749-b25b-577325d23704 req-462a5fb7-12e8-4619-8443-7a6c4f70f7c0 service nova] Acquiring lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.169126] env[65121]: DEBUG oslo_concurrency.lockutils [req-ddd1ca7f-814a-4749-b25b-577325d23704 req-462a5fb7-12e8-4619-8443-7a6c4f70f7c0 service nova] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.169326] env[65121]: DEBUG oslo_concurrency.lockutils [req-ddd1ca7f-814a-4749-b25b-577325d23704 req-462a5fb7-12e8-4619-8443-7a6c4f70f7c0 service nova] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.169595] env[65121]: DEBUG nova.compute.manager [req-ddd1ca7f-814a-4749-b25b-577325d23704 req-462a5fb7-12e8-4619-8443-7a6c4f70f7c0 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] No waiting events found dispatching network-vif-plugged-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1142.169832] env[65121]: WARNING nova.compute.manager [req-ddd1ca7f-814a-4749-b25b-577325d23704 req-462a5fb7-12e8-4619-8443-7a6c4f70f7c0 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received unexpected event network-vif-plugged-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf for instance with vm_state active and task_state None. [ 1142.173924] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1142.174472] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-510680b6-9190-4e4f-ab40-3786bd2f497a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.182396] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1142.182396] env[65121]: value = "task-5107351" [ 1142.182396] env[65121]: _type = "Task" [ 1142.182396] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.197742] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107351, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.234173] env[65121]: DEBUG nova.network.neutron [-] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1142.384345] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.384345] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.384560] env[65121]: DEBUG nova.network.neutron [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1142.403875] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: bad70910-58a1-4e35-8d0c-3bd7c3a30abf] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1142.535106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.537984] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.387s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.539622] env[65121]: INFO nova.compute.claims [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.651519] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-90ab2917-9314-47d1-8b0d-2d9aaba4b421 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.662120] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea23161a-c167-43f0-bd2b-b31bfffa1451 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.681238] env[65121]: INFO nova.scheduler.client.report [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Deleted allocations for instance 714930d8-3ec9-4aa8-831a-a9f28c365545 [ 1142.709711] env[65121]: DEBUG nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1142.712448] env[65121]: DEBUG nova.compute.manager [req-6a811323-6328-414b-a195-8b950ded5b97 req-7a40f0cb-a096-4821-8af9-8ea7f048c89a service nova] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Detach interface failed, port_id=e31cbeae-05bc-416d-9df3-10f09d947ba4, reason: Instance 6fe86deb-0903-4769-a05c-b7d0acec0103 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1142.721209] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107351, 'name': PowerOffVM_Task, 'duration_secs': 0.236822} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.721209] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1142.721377] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance '84ae6061-a962-4a74-8661-9718cc4c5346' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1142.735855] env[65121]: INFO nova.compute.manager [-] [instance: 6fe86deb-0903-4769-a05c-b7d0acec0103] Took 1.58 seconds to deallocate network for instance. [ 1142.742995] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1142.742995] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1142.743230] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1142.743263] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1142.743439] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1142.743519] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1142.743713] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1142.743856] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1142.744016] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1142.744809] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1142.744809] env[65121]: DEBUG nova.virt.hardware [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1142.745324] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5869fc-0f03-45da-a4bd-1cb9a67a3e4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.755728] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493c878e-331c-4ac9-81f1-b9f6bccb6810 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.887636] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.888035] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.909785] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: efb98d37-4162-4249-9f85-008d4537db87] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1142.943319] env[65121]: WARNING nova.network.neutron [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] 5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd already exists in list: networks containing: ['5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd']. ignoring it [ 1142.983479] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.984122] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.094377] env[65121]: WARNING neutronclient.v2_0.client [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.095258] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.095619] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.216926] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a6f35e7c-3119-4da1-aa39-fed183229dfc tempest-InstanceActionsNegativeTestJSON-1526118371 tempest-InstanceActionsNegativeTestJSON-1526118371-project-member] Lock "714930d8-3ec9-4aa8-831a-a9f28c365545" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.885s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.227908] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1143.228168] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1143.228367] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1143.228559] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1143.228669] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1143.228779] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1143.228973] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1143.229214] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1143.229392] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1143.229647] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1143.229730] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1143.236210] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06617164-2e17-4b71-a6c4-0c128ab3597f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.249220] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.256706] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1143.256706] env[65121]: value = "task-5107352" [ 1143.256706] env[65121]: _type = "Task" [ 1143.256706] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.267777] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107352, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.268830] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.269195] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.346715] env[65121]: WARNING neutronclient.v2_0.client [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.346802] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.347367] env[65121]: WARNING openstack [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.414877] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 564159fe-6ecd-4276-8d65-cbe25859493a] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1143.653992] env[65121]: DEBUG nova.network.neutron [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Successfully updated port: 91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1143.764992] env[65121]: DEBUG nova.network.neutron [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "address": "fa:16:3e:16:2f:1f", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6ae8a69-75", "ovs_interfaceid": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1143.769869] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107352, 'name': ReconfigVM_Task, 'duration_secs': 0.377416} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.773460] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance '84ae6061-a962-4a74-8661-9718cc4c5346' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1143.834360] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732b1a8f-5567-43fd-8590-bc8f08b66df4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.844180] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818d3726-0d55-4c64-a6ff-75f8e48eff44 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.875205] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffa145a-acb0-472e-b397-f4f36d9615c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.883363] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2faaaa-f18a-4f0e-8ade-4f0ba7c4061f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.899302] env[65121]: DEBUG nova.compute.provider_tree [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.918720] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 17f34d32-e690-40d8-99a4-acf53fa859c6] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1144.159268] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.159502] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.159691] env[65121]: DEBUG nova.network.neutron [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1144.270897] env[65121]: DEBUG nova.compute.manager [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-changed-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1144.271259] env[65121]: DEBUG nova.compute.manager [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing instance network info cache due to event network-changed-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1144.271525] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.272090] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.272812] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.272996] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.273635] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.273854] env[65121]: DEBUG nova.network.neutron [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing network info cache for port f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1144.275633] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70603c9-9440-464b-974a-573217c51f62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.281180] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1144.281447] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1144.281635] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1144.281847] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1144.282034] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1144.282216] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1144.282451] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.282639] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1144.282837] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1144.283048] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1144.283274] env[65121]: DEBUG nova.virt.hardware [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1144.288646] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1144.289786] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1dd3ddb9-e6a8-4a9a-ac69-bea238f1e8c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.320148] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1144.320404] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1144.320551] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1144.320723] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1144.320859] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1144.320996] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1144.321205] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.321359] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1144.321515] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1144.321667] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1144.321828] env[65121]: DEBUG nova.virt.hardware [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1144.328435] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Reconfiguring VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1144.330917] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b89557ed-be02-4ee6-8348-4d648bd7e170 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.344437] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1144.344437] env[65121]: value = "task-5107353" [ 1144.344437] env[65121]: _type = "Task" [ 1144.344437] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.351373] env[65121]: DEBUG oslo_vmware.api [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1144.351373] env[65121]: value = "task-5107354" [ 1144.351373] env[65121]: _type = "Task" [ 1144.351373] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.357736] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107353, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.364187] env[65121]: DEBUG oslo_vmware.api [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107354, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.403292] env[65121]: DEBUG nova.scheduler.client.report [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1144.422206] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: e46c88ed-2c2e-432d-987d-0e4c99c28e42] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1144.663248] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.663248] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.744922] env[65121]: DEBUG nova.network.neutron [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1144.780682] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.781114] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.787842] env[65121]: WARNING neutronclient.v2_0.client [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.788474] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.788796] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.860021] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107353, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.866253] env[65121]: DEBUG oslo_vmware.api [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107354, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.878113] env[65121]: WARNING neutronclient.v2_0.client [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.878997] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.879441] env[65121]: WARNING openstack [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.909626] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.910188] env[65121]: DEBUG nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1144.914416] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.978s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.927310] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.927542] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.937533] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 23d6f632-8918-46af-b239-08a9615dfbec] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1145.004869] env[65121]: DEBUG nova.network.neutron [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91969a04-4a", "ovs_interfaceid": "91969a04-4a8e-4334-829f-3e3ae6828511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1145.038169] env[65121]: WARNING neutronclient.v2_0.client [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.038943] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.039388] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.182533] env[65121]: DEBUG nova.network.neutron [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updated VIF entry in instance network info cache for port f6ae8a69-7565-42c4-96fe-1fbd5e038ebf. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1145.182993] env[65121]: DEBUG nova.network.neutron [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "address": "fa:16:3e:16:2f:1f", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6ae8a69-75", "ovs_interfaceid": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1145.355238] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107353, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.364087] env[65121]: DEBUG oslo_vmware.api [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107354, 'name': ReconfigVM_Task, 'duration_secs': 0.823197} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.364524] env[65121]: WARNING neutronclient.v2_0.client [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.364768] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.364977] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Reconfigured VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1145.416504] env[65121]: DEBUG nova.compute.utils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1145.418039] env[65121]: DEBUG nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1145.418918] env[65121]: DEBUG nova.network.neutron [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1145.418918] env[65121]: WARNING neutronclient.v2_0.client [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.419070] env[65121]: WARNING neutronclient.v2_0.client [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.419597] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.419930] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.446881] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 90cb98b1-9520-4caf-b6c4-80abaa7a9221] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1145.508605] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.508861] env[65121]: DEBUG nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Instance network_info: |[{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91969a04-4a", "ovs_interfaceid": "91969a04-4a8e-4334-829f-3e3ae6828511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1145.510029] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:16:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a64108f9-df0a-4feb-bbb5-97f5841c356c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91969a04-4a8e-4334-829f-3e3ae6828511', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1145.517715] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1145.521297] env[65121]: DEBUG nova.policy [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26beb81d28f44cc8932ea8e87a4cd2c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0630960dcbf44781be05184565d81932', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1145.526583] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1145.528809] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf4793d0-c82b-4ad5-8aaa-9791762138f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.553427] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1145.553427] env[65121]: value = "task-5107355" [ 1145.553427] env[65121]: _type = "Task" [ 1145.553427] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.562321] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107355, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.685790] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.686285] env[65121]: DEBUG nova.compute.manager [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-vif-plugged-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1145.686488] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1145.686681] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1145.686830] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.686986] env[65121]: DEBUG nova.compute.manager [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] No waiting events found dispatching network-vif-plugged-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1145.687162] env[65121]: WARNING nova.compute.manager [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received unexpected event network-vif-plugged-91969a04-4a8e-4334-829f-3e3ae6828511 for instance with vm_state building and task_state spawning. [ 1145.687316] env[65121]: DEBUG nova.compute.manager [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1145.687465] env[65121]: DEBUG nova.compute.manager [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing instance network info cache due to event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1145.687720] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Acquiring lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.687872] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Acquired lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.688043] env[65121]: DEBUG nova.network.neutron [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1145.719621] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2268b53a-1354-4b39-ad68-b41ea92e3886 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.728953] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92587ee-44f6-4a28-8c7a-5d6afe2e7e08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.763625] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc78996a-a400-489c-a5fc-3decc957b883 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.772514] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9b1cdc-8df3-46c4-b8be-d4f23f4ebe6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.787583] env[65121]: DEBUG nova.compute.provider_tree [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.857768] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107353, 'name': ReconfigVM_Task, 'duration_secs': 1.213297} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.857879] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1145.859124] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a543ee89-fc18-4936-96ed-60660185b960 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.880322] env[65121]: DEBUG oslo_concurrency.lockutils [None req-18947a46-42a0-48d2-9c5f-1fc54e78ffa0 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.786s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.890422] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1145.891800] env[65121]: DEBUG nova.network.neutron [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Successfully created port: d86b7885-6e5e-45ec-b64e-19ae24268b82 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1145.895087] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f143275e-3d92-4c59-b144-78b39f400b11 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.919037] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1145.919037] env[65121]: value = "task-5107356" [ 1145.919037] env[65121]: _type = "Task" [ 1145.919037] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.927614] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107356, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.930480] env[65121]: DEBUG nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1145.954839] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 59d6005d-6caf-4898-b791-70d9c015cdb4] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1146.065598] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107355, 'name': CreateVM_Task, 'duration_secs': 0.399068} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.066322] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1146.066548] env[65121]: WARNING neutronclient.v2_0.client [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.066915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.067164] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.067764] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1146.067847] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da924a7a-918d-4ce4-a27b-aa4bf1cd05b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.073428] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1146.073428] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52137328-91f6-3d43-d0bf-f2a2c7a1dc76" [ 1146.073428] env[65121]: _type = "Task" [ 1146.073428] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.083279] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52137328-91f6-3d43-d0bf-f2a2c7a1dc76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.192879] env[65121]: WARNING neutronclient.v2_0.client [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.194054] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.194673] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.291066] env[65121]: DEBUG nova.scheduler.client.report [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1146.431183] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107356, 'name': ReconfigVM_Task, 'duration_secs': 0.334131} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.431554] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1146.431874] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance '84ae6061-a962-4a74-8661-9718cc4c5346' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1146.458744] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: ac001ee6-f601-48f7-af0d-42c2387f3c98] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1146.591023] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52137328-91f6-3d43-d0bf-f2a2c7a1dc76, 'name': SearchDatastore_Task, 'duration_secs': 0.013262} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.591023] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1146.591023] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1146.591023] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.591023] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.591023] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1146.591023] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a2b1b9e-6476-48e4-8949-1f2e7fa687be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.600581] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1146.603017] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1146.603017] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9a94f95-306d-484d-b13a-01455aed9dd7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.611461] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1146.611461] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525383a8-3910-1ebe-eb12-22c781e8f601" [ 1146.611461] env[65121]: _type = "Task" [ 1146.611461] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.622634] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525383a8-3910-1ebe-eb12-22c781e8f601, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.943880] env[65121]: DEBUG nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1146.946634] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d977ba4a-dccd-4347-9ca3-03a95d9e473a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.971306] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 49b7e6c3-9106-4b57-ae44-85b5d4474089] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1146.975815] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40c95ff-5fec-4f7d-8628-bfc18752a538 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.986137] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1146.986379] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1146.986593] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1146.987706] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1146.987706] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1146.987706] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1146.987706] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.987706] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1146.987706] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1146.987957] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1146.987957] env[65121]: DEBUG nova.virt.hardware [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1147.006689] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e66934-5d6f-4f39-8933-2307fceba366 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.009565] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance '84ae6061-a962-4a74-8661-9718cc4c5346' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1147.019496] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0038dff-f0d8-4871-9eaa-f6b345c4bd39 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.120692] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525383a8-3910-1ebe-eb12-22c781e8f601, 'name': SearchDatastore_Task, 'duration_secs': 0.013599} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.121533] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08c0de7c-bb06-4caf-abd7-a9fc28a8e586 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.127891] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1147.127891] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522aa686-6bab-a1f5-494c-5455c4a97b46" [ 1147.127891] env[65121]: _type = "Task" [ 1147.127891] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.139082] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522aa686-6bab-a1f5-494c-5455c4a97b46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.306062] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.391s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.309977] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.369s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.312152] env[65121]: INFO nova.compute.claims [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1147.479870] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: c52a1269-bb34-4ef6-ab8d-78df3e9b1d39] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1147.495612] env[65121]: DEBUG nova.network.neutron [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Successfully updated port: d86b7885-6e5e-45ec-b64e-19ae24268b82 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1147.643317] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522aa686-6bab-a1f5-494c-5455c4a97b46, 'name': SearchDatastore_Task, 'duration_secs': 0.025536} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.643317] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.643317] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1147.643317] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09d9bae8-a56c-4c46-b785-c52b9add96ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.650557] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1147.650557] env[65121]: value = "task-5107357" [ 1147.650557] env[65121]: _type = "Task" [ 1147.650557] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.659505] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107357, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.671617] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.672016] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.804778] env[65121]: WARNING neutronclient.v2_0.client [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.805513] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.806075] env[65121]: WARNING openstack [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.911539] env[65121]: INFO nova.scheduler.client.report [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted allocation for migration f186477a-a599-4c02-99d7-f463c3b63131 [ 1147.945282] env[65121]: DEBUG nova.network.neutron [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updated VIF entry in instance network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1147.945282] env[65121]: DEBUG nova.network.neutron [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91969a04-4a", "ovs_interfaceid": "91969a04-4a8e-4334-829f-3e3ae6828511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1147.983489] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: b2ebce1d-a8d8-4a71-997b-39d09ca373fe] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1147.993293] env[65121]: DEBUG nova.compute.manager [req-ce1d91b5-b3aa-4eb6-bdac-8515251f090d req-023f1590-4ff4-40d3-a6b3-63a114341cda service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Received event network-vif-plugged-d86b7885-6e5e-45ec-b64e-19ae24268b82 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1147.993401] env[65121]: DEBUG oslo_concurrency.lockutils [req-ce1d91b5-b3aa-4eb6-bdac-8515251f090d req-023f1590-4ff4-40d3-a6b3-63a114341cda service nova] Acquiring lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1147.993516] env[65121]: DEBUG oslo_concurrency.lockutils [req-ce1d91b5-b3aa-4eb6-bdac-8515251f090d req-023f1590-4ff4-40d3-a6b3-63a114341cda service nova] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.993842] env[65121]: DEBUG oslo_concurrency.lockutils [req-ce1d91b5-b3aa-4eb6-bdac-8515251f090d req-023f1590-4ff4-40d3-a6b3-63a114341cda service nova] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.993842] env[65121]: DEBUG nova.compute.manager [req-ce1d91b5-b3aa-4eb6-bdac-8515251f090d req-023f1590-4ff4-40d3-a6b3-63a114341cda service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] No waiting events found dispatching network-vif-plugged-d86b7885-6e5e-45ec-b64e-19ae24268b82 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1147.993985] env[65121]: WARNING nova.compute.manager [req-ce1d91b5-b3aa-4eb6-bdac-8515251f090d req-023f1590-4ff4-40d3-a6b3-63a114341cda service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Received unexpected event network-vif-plugged-d86b7885-6e5e-45ec-b64e-19ae24268b82 for instance with vm_state building and task_state spawning. [ 1147.998482] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.998482] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1147.998636] env[65121]: DEBUG nova.network.neutron [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1148.164337] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107357, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.423042] env[65121]: DEBUG oslo_concurrency.lockutils [None req-83d901ff-4683-4674-997e-f97b3b6b355c tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.552s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.448017] env[65121]: DEBUG oslo_concurrency.lockutils [req-2ddec842-b2dd-428b-841e-14926e8d2677 req-681039a4-a723-46e8-a9e0-23f35fd205a6 service nova] Releasing lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1148.490527] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 600d5554-f52a-48ca-941d-1a755d086823] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1148.505939] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.505939] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.559480] env[65121]: DEBUG nova.network.neutron [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1148.599891] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.599891] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.611040] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d3fe22-8f69-486c-88d7-407e7b6bedd7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.619284] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c73d55-65cb-497c-9cd5-cfe8abb51d3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.655511] env[65121]: WARNING neutronclient.v2_0.client [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.667030] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db2837a-f7c6-4e44-aad7-97b5119e5418 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.680923] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0120430c-f0e2-4fba-80a7-9c010a5a7fdb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.685542] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.685714] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.687615] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107357, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.687057} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.687615] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1148.687761] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1148.688392] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc543f41-6ea8-4350-af18-f679de6a41f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.701304] env[65121]: DEBUG nova.compute.provider_tree [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.709052] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1148.709052] env[65121]: value = "task-5107358" [ 1148.709052] env[65121]: _type = "Task" [ 1148.709052] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.713399] env[65121]: DEBUG nova.network.neutron [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Port 2acd5dbc-a08c-4ebd-922b-284294369a33 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1148.720118] env[65121]: WARNING neutronclient.v2_0.client [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.720118] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.720118] env[65121]: WARNING openstack [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.740269] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107358, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.833496] env[65121]: DEBUG nova.network.neutron [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance_info_cache with network_info: [{"id": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "address": "fa:16:3e:22:e8:da", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd86b7885-6e", "ovs_interfaceid": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1148.867172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.867172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.867172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.867172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.867172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.869979] env[65121]: INFO nova.compute.manager [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Terminating instance [ 1148.969685] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.970153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.970153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.970580] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.970580] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.972605] env[65121]: INFO nova.compute.manager [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Terminating instance [ 1148.995491] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 2e676b64-181e-4fee-8120-05cb49b5bb4e] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1149.174873] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.175126] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.201441] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.201708] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1149.202770] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5108ba-8e6e-4d7c-b932-3fff31920c4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.207433] env[65121]: DEBUG nova.scheduler.client.report [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1149.234114] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c7548f-a8dd-4f77-84b9-a8295ebce8f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.249597] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107358, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082434} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.251679] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1149.253244] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b44816d-5751-414f-a887-546811b1d18d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.275303] env[65121]: WARNING neutronclient.v2_0.client [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1149.281282] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Reconfiguring VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1149.282336] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b71fbd36-68b3-4b63-b4a2-39cac40f0b2e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.318638] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.320059] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a232dac-ff7b-4780-a8b1-6b29a3fc6a2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.338582] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1149.338928] env[65121]: DEBUG nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Instance network_info: |[{"id": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "address": "fa:16:3e:22:e8:da", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd86b7885-6e", "ovs_interfaceid": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1149.340044] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1149.340044] env[65121]: value = "task-5107359" [ 1149.340044] env[65121]: _type = "Task" [ 1149.340044] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.340044] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:e8:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd86b7885-6e5e-45ec-b64e-19ae24268b82', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1149.348181] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1149.349409] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1149.353968] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3ed5d6a-4fd5-4d24-8ad0-f28e8e2037b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.369753] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1149.369753] env[65121]: value = "task-5107360" [ 1149.369753] env[65121]: _type = "Task" [ 1149.369753] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.380213] env[65121]: DEBUG nova.compute.manager [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1149.380479] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1149.380817] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.381106] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1149.381106] env[65121]: value = "task-5107361" [ 1149.381106] env[65121]: _type = "Task" [ 1149.381106] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.382557] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c768ae7-bb13-40f4-b6c5-7ffb059789fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.397461] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107360, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.404472] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1149.406523] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ff46da7-85aa-437b-a295-a009b01f766a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.407845] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107361, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.414924] env[65121]: DEBUG oslo_vmware.api [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1149.414924] env[65121]: value = "task-5107362" [ 1149.414924] env[65121]: _type = "Task" [ 1149.414924] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.424754] env[65121]: DEBUG oslo_vmware.api [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.477102] env[65121]: DEBUG nova.compute.manager [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1149.477942] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1149.479037] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555c01a4-d5bd-431d-be49-4047c66ffad4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.490500] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1149.490900] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88a5ed7d-fc70-43ff-ae6e-1bd815732f93 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.500213] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: df5abcb5-583f-4b28-a074-3a3221d74d87] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1149.503250] env[65121]: DEBUG oslo_vmware.api [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1149.503250] env[65121]: value = "task-5107363" [ 1149.503250] env[65121]: _type = "Task" [ 1149.503250] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.517335] env[65121]: DEBUG oslo_vmware.api [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107363, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.681466] env[65121]: DEBUG nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1149.714038] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.714669] env[65121]: DEBUG nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1149.717606] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.371s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.717872] env[65121]: DEBUG nova.objects.instance [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'resources' on Instance uuid 59ceb0fd-1fb5-4c90-963d-fe76b9740d29 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.770811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9475863f-86d5-44e6-ac19-93461ab87743" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.770811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9475863f-86d5-44e6-ac19-93461ab87743" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.778611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.778611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.778611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.797376] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.799392] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.861573] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.884345] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107360, 'name': ReconfigVM_Task, 'duration_secs': 0.32531} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.887229] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1149.887229] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a6024c7-c680-4513-8ea9-cfe9a4663865 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.898398] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107361, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.899863] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1149.899863] env[65121]: value = "task-5107364" [ 1149.899863] env[65121]: _type = "Task" [ 1149.899863] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.908356] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107364, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.925278] env[65121]: DEBUG oslo_vmware.api [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107362, 'name': PowerOffVM_Task, 'duration_secs': 0.303954} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.925543] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1149.925723] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1149.925941] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6117cf82-d55d-488f-bcb5-c598868d39e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.999923] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1150.000268] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1150.000368] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleting the datastore file [datastore2] 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1150.000662] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b8c0e5f-f6d6-4586-909e-32dcbd2be25d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.010088] env[65121]: DEBUG oslo_vmware.api [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for the task: (returnval){ [ 1150.010088] env[65121]: value = "task-5107366" [ 1150.010088] env[65121]: _type = "Task" [ 1150.010088] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.011289] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 30b01b2b-2cbf-4c34-86cd-529a95a9c3ec] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1150.022070] env[65121]: DEBUG oslo_vmware.api [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107363, 'name': PowerOffVM_Task, 'duration_secs': 0.228692} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.022860] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1150.023094] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1150.024476] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7eca274-76c9-4016-80c3-11c4bdd050a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.028860] env[65121]: DEBUG oslo_vmware.api [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.031835] env[65121]: DEBUG nova.compute.manager [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Received event network-changed-d86b7885-6e5e-45ec-b64e-19ae24268b82 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1150.032114] env[65121]: DEBUG nova.compute.manager [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Refreshing instance network info cache due to event network-changed-d86b7885-6e5e-45ec-b64e-19ae24268b82. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1150.032320] env[65121]: DEBUG oslo_concurrency.lockutils [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Acquiring lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.032378] env[65121]: DEBUG oslo_concurrency.lockutils [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Acquired lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.032562] env[65121]: DEBUG nova.network.neutron [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Refreshing network info cache for port d86b7885-6e5e-45ec-b64e-19ae24268b82 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1150.214775] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1150.220337] env[65121]: DEBUG nova.compute.utils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1150.222374] env[65121]: DEBUG nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1150.222614] env[65121]: DEBUG nova.network.neutron [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1150.223345] env[65121]: WARNING neutronclient.v2_0.client [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.223658] env[65121]: WARNING neutronclient.v2_0.client [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.224244] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.225039] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.233130] env[65121]: DEBUG nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1150.280453] env[65121]: DEBUG nova.policy [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1c51fc43b7f44ce9b023d8cb769e029', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89110618b6ef4ce68b2ca53e7e168139', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1150.283048] env[65121]: DEBUG nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1150.303966] env[65121]: DEBUG nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1150.366318] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.396600] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107361, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.203171] env[65121]: DEBUG nova.network.neutron [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Successfully created port: a2830632-dca1-42e0-be06-922a62ab2307 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1151.206359] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 274d0ccd-c707-4a68-b280-16de2bc74d73] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1151.211108] env[65121]: WARNING neutronclient.v2_0.client [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.211108] env[65121]: WARNING openstack [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.211108] env[65121]: WARNING openstack [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.221638] env[65121]: WARNING neutronclient.v2_0.client [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.225241] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1151.228104] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1151.228104] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleting the datastore file [datastore1] 9f9118f1-c708-469f-839e-b3c4ef52f38f {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1151.238750] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25b6241e-4d60-49b6-a9b8-f7159c01af3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.249024] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107364, 'name': Rename_Task, 'duration_secs': 0.163663} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.249024] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1151.251422] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93c06b62-2ce2-49b2-966b-0858eadfc726 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.253486] env[65121]: DEBUG oslo_vmware.api [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Task: {'id': task-5107366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186199} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.264151] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1151.264377] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1151.264716] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1151.265040] env[65121]: INFO nova.compute.manager [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Took 1.88 seconds to destroy the instance on the hypervisor. [ 1151.265318] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1151.266827] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107361, 'name': CreateVM_Task, 'duration_secs': 1.487817} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.266827] env[65121]: DEBUG oslo_vmware.api [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1151.266827] env[65121]: value = "task-5107368" [ 1151.266827] env[65121]: _type = "Task" [ 1151.266827] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.266827] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.267381] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.268283] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.270866] env[65121]: DEBUG nova.compute.manager [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1151.270866] env[65121]: DEBUG nova.network.neutron [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1151.270866] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.270866] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.270866] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.278921] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1151.281626] env[65121]: WARNING neutronclient.v2_0.client [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.282189] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.282189] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.282999] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1151.286255] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1151.286255] env[65121]: value = "task-5107369" [ 1151.286255] env[65121]: _type = "Task" [ 1151.286255] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.290799] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db8914f3-2e48-4360-8356-e2f2245131a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.301500] env[65121]: DEBUG oslo_vmware.api [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.305213] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1151.305213] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52731be2-cf8b-d042-0c5a-1a148d0c6810" [ 1151.305213] env[65121]: _type = "Task" [ 1151.305213] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.312761] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107369, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.323322] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52731be2-cf8b-d042-0c5a-1a148d0c6810, 'name': SearchDatastore_Task, 'duration_secs': 0.012364} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.324617] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.324832] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.325191] env[65121]: DEBUG nova.network.neutron [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1151.326974] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.326974] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1151.326974] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.327249] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.327370] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1151.330692] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a4dae5a-9df2-441b-be06-c85c7c008890 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.341151] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1151.341282] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1151.342286] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72b61bd9-542e-4129-8146-70170eb0b83f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.351556] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1151.351556] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a03edc-79e9-c654-408e-e97b02cc5a83" [ 1151.351556] env[65121]: _type = "Task" [ 1151.351556] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.355419] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.366383] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a03edc-79e9-c654-408e-e97b02cc5a83, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.367268] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-264a7c0a-7844-4ccf-b04d-a60bae99e3a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.378228] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1151.378228] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a2a4fb-ff48-1337-d2c5-81f8871df139" [ 1151.378228] env[65121]: _type = "Task" [ 1151.378228] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.385363] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a2a4fb-ff48-1337-d2c5-81f8871df139, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.432171] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb41512-2871-404f-9523-de4de03999ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.441399] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa9241d-7d60-43b7-a485-24ad415eeaf3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.480220] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2e9472-2f7c-4256-b39e-bd4f55ac5521 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.489413] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29a3c8b-c237-4699-8cec-a51f82819434 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.504956] env[65121]: DEBUG nova.compute.provider_tree [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.702482] env[65121]: WARNING openstack [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.702895] env[65121]: WARNING openstack [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.732305] env[65121]: DEBUG nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1151.736360] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 68e3e4e6-6ca1-4f76-a4af-5112a2042c61] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1151.741541] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.766233] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1151.766468] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1151.766642] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1151.766809] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1151.766944] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1151.767132] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1151.767405] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1151.767567] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1151.767725] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1151.767877] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1151.768285] env[65121]: DEBUG nova.virt.hardware [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1151.769020] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a93e9e4-6686-4af8-b4c7-f23976273ee1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.779368] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691d4a2d-6374-40f8-a12a-b612f4682dc5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.807838] env[65121]: DEBUG oslo_vmware.api [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234555} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.808101] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1151.808304] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1151.808492] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1151.808660] env[65121]: INFO nova.compute.manager [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Took 2.33 seconds to destroy the instance on the hypervisor. [ 1151.808891] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1151.809570] env[65121]: DEBUG nova.compute.manager [-] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1151.809658] env[65121]: DEBUG nova.network.neutron [-] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1151.809948] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.810495] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.811282] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.821260] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107369, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.833303] env[65121]: WARNING neutronclient.v2_0.client [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.833935] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.834296] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.890963] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a2a4fb-ff48-1337-d2c5-81f8871df139, 'name': SearchDatastore_Task, 'duration_secs': 0.018383} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.891312] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.891584] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 9a248b47-a34b-4a98-baf0-aa971b5aca94/9a248b47-a34b-4a98-baf0-aa971b5aca94.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1151.891850] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd67455a-bf75-4308-a397-ca5c016296eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.900579] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1151.900579] env[65121]: value = "task-5107370" [ 1151.900579] env[65121]: _type = "Task" [ 1151.900579] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.909487] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.910594] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.966526] env[65121]: WARNING neutronclient.v2_0.client [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.967320] env[65121]: WARNING openstack [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.967665] env[65121]: WARNING openstack [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.009142] env[65121]: DEBUG nova.scheduler.client.report [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1152.187182] env[65121]: DEBUG nova.network.neutron [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1152.238161] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.243969] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 1b6ad893-e013-4fd5-a829-535d1ea6e001] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1152.305955] env[65121]: DEBUG oslo_vmware.api [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107369, 'name': PowerOnVM_Task, 'duration_secs': 0.545961} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.307213] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1152.307652] env[65121]: INFO nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Took 9.60 seconds to spawn the instance on the hypervisor. [ 1152.307819] env[65121]: DEBUG nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1152.314076] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb66377d-67e0-4584-abf8-a1917f30910f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.418266] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107370, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510992} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.418705] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 9a248b47-a34b-4a98-baf0-aa971b5aca94/9a248b47-a34b-4a98-baf0-aa971b5aca94.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1152.418922] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1152.419224] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf83917f-239a-4b18-a222-598e4b76df62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.430641] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1152.430641] env[65121]: value = "task-5107371" [ 1152.430641] env[65121]: _type = "Task" [ 1152.430641] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.440146] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.514168] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.796s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.516574] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.267s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.516821] env[65121]: DEBUG nova.objects.instance [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'resources' on Instance uuid 6fe86deb-0903-4769-a05c-b7d0acec0103 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.535555] env[65121]: INFO nova.scheduler.client.report [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted allocations for instance 59ceb0fd-1fb5-4c90-963d-fe76b9740d29 [ 1152.692459] env[65121]: INFO nova.compute.manager [-] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Took 1.42 seconds to deallocate network for instance. [ 1152.717401] env[65121]: DEBUG nova.network.neutron [-] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1152.738643] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.749417] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 8a23da27-ce6c-4453-9036-65eeeda3ce0d] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1152.836192] env[65121]: INFO nova.compute.manager [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Took 19.35 seconds to build instance. [ 1152.846102] env[65121]: DEBUG nova.network.neutron [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Successfully updated port: a2830632-dca1-42e0-be06-922a62ab2307 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1152.883571] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.883944] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.894107] env[65121]: DEBUG nova.network.neutron [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updated VIF entry in instance network info cache for port d86b7885-6e5e-45ec-b64e-19ae24268b82. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1152.894454] env[65121]: DEBUG nova.network.neutron [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance_info_cache with network_info: [{"id": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "address": "fa:16:3e:22:e8:da", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd86b7885-6e", "ovs_interfaceid": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1152.926596] env[65121]: DEBUG nova.compute.manager [req-6c5ea28b-1292-46eb-b66f-4c0130f84393 req-6f08bd6b-fe77-456c-a257-9eb410743bd6 service nova] [instance: 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731] Received event network-vif-deleted-04d47dd2-7f48-48e4-aed4-9d693168b8ca {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1152.942624] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24591} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.946626] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1152.947357] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd66e76-fcdd-4512-b66a-b4a6ef209334 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.952746] env[65121]: WARNING neutronclient.v2_0.client [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.953399] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.953742] env[65121]: WARNING openstack [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.981315] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 9a248b47-a34b-4a98-baf0-aa971b5aca94/9a248b47-a34b-4a98-baf0-aa971b5aca94.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.982297] env[65121]: DEBUG nova.compute.manager [req-ca74aed5-f98f-49e1-b52d-4c630566d29c req-1a7d665e-b882-48f3-913f-07b4ae658046 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Received event network-vif-plugged-a2830632-dca1-42e0-be06-922a62ab2307 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1152.982485] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca74aed5-f98f-49e1-b52d-4c630566d29c req-1a7d665e-b882-48f3-913f-07b4ae658046 service nova] Acquiring lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.982677] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca74aed5-f98f-49e1-b52d-4c630566d29c req-1a7d665e-b882-48f3-913f-07b4ae658046 service nova] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.982829] env[65121]: DEBUG oslo_concurrency.lockutils [req-ca74aed5-f98f-49e1-b52d-4c630566d29c req-1a7d665e-b882-48f3-913f-07b4ae658046 service nova] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.982977] env[65121]: DEBUG nova.compute.manager [req-ca74aed5-f98f-49e1-b52d-4c630566d29c req-1a7d665e-b882-48f3-913f-07b4ae658046 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] No waiting events found dispatching network-vif-plugged-a2830632-dca1-42e0-be06-922a62ab2307 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1152.983143] env[65121]: WARNING nova.compute.manager [req-ca74aed5-f98f-49e1-b52d-4c630566d29c req-1a7d665e-b882-48f3-913f-07b4ae658046 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Received unexpected event network-vif-plugged-a2830632-dca1-42e0-be06-922a62ab2307 for instance with vm_state building and task_state spawning. [ 1152.983944] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-360c6b81-d74c-4745-a66b-01a231dd2cdf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.005037] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1153.005037] env[65121]: value = "task-5107372" [ 1153.005037] env[65121]: _type = "Task" [ 1153.005037] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.018436] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107372, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.049912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a409f367-8950-4e8d-aa8c-baa121d73be7 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "59ceb0fd-1fb5-4c90-963d-fe76b9740d29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.442s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.063610] env[65121]: DEBUG nova.network.neutron [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1153.199858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.220242] env[65121]: INFO nova.compute.manager [-] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Took 1.41 seconds to deallocate network for instance. [ 1153.240633] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.253939] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 7fcab9cb-9a0f-4a50-9244-687889443a31] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1153.329112] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6d2c41-4ba4-4d9d-aa57-dc9e77dd12ab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.337964] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090d8d07-ac16-4882-a5c7-97cbb287b21d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.342039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-af004497-5cde-4bbc-9046-b719afd508e0 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.862s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.375712] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "refresh_cache-f85ae5c9-ce3f-4ade-b708-4eae79485b20" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.375885] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "refresh_cache-f85ae5c9-ce3f-4ade-b708-4eae79485b20" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.376103] env[65121]: DEBUG nova.network.neutron [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1153.378916] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d81bdb-b478-4390-ba2a-096e5e19d644 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.388812] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992c6835-8ce3-4fd0-a9f1-a0d55bfcdea8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.405841] env[65121]: DEBUG oslo_concurrency.lockutils [req-c8394ff3-20e3-4095-8775-f9540c9b8f0b req-26243453-4058-4d81-8c40-84210f4fc4c0 service nova] Releasing lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.406892] env[65121]: DEBUG nova.compute.provider_tree [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.519294] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107372, 'name': ReconfigVM_Task, 'duration_secs': 0.342223} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.519580] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 9a248b47-a34b-4a98-baf0-aa971b5aca94/9a248b47-a34b-4a98-baf0-aa971b5aca94.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1153.520553] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01c7caf1-f728-4983-a103-5c5e4526def3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.528850] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1153.528850] env[65121]: value = "task-5107373" [ 1153.528850] env[65121]: _type = "Task" [ 1153.528850] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.537874] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107373, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.574725] env[65121]: DEBUG oslo_concurrency.lockutils [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.727590] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.737540] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.756220] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 93721179-407a-43d9-b0bf-157433a09519] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1153.880087] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.880525] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.911519] env[65121]: DEBUG nova.scheduler.client.report [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1153.922647] env[65121]: DEBUG nova.network.neutron [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1154.011034] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.011034] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.040659] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107373, 'name': Rename_Task, 'duration_secs': 0.163377} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.041040] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1154.041639] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01e06af6-fa30-4471-add4-97efa8817ad0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.048570] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1154.048570] env[65121]: value = "task-5107374" [ 1154.048570] env[65121]: _type = "Task" [ 1154.048570] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.066610] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.085220] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88ac6a7-5a7e-40f6-a560-12cb31fc2cde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.093631] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e160496-7b77-4728-9252-3bea7a047980 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.104256] env[65121]: WARNING neutronclient.v2_0.client [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.104957] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.105318] env[65121]: WARNING openstack [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.238544] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.248854] env[65121]: DEBUG nova.network.neutron [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Updating instance_info_cache with network_info: [{"id": "a2830632-dca1-42e0-be06-922a62ab2307", "address": "fa:16:3e:4f:9c:29", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2830632-dc", "ovs_interfaceid": "a2830632-dca1-42e0-be06-922a62ab2307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.261261] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 99b1ff9c-38b7-4947-81bb-d200a3b2a09a] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1154.417107] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.420034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.205s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.421666] env[65121]: INFO nova.compute.claims [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1154.452981] env[65121]: INFO nova.scheduler.client.report [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Deleted allocations for instance 6fe86deb-0903-4769-a05c-b7d0acec0103 [ 1154.560814] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107374, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.740219] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.752606] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "refresh_cache-f85ae5c9-ce3f-4ade-b708-4eae79485b20" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1154.753011] env[65121]: DEBUG nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Instance network_info: |[{"id": "a2830632-dca1-42e0-be06-922a62ab2307", "address": "fa:16:3e:4f:9c:29", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2830632-dc", "ovs_interfaceid": "a2830632-dca1-42e0-be06-922a62ab2307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1154.753505] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:9c:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b80dd748-3d7e-4a23-a38d-9e79a3881452', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a2830632-dca1-42e0-be06-922a62ab2307', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1154.769511] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1154.770566] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 5d352184-7074-4fab-b4ac-cc11a6b936b1] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1154.773488] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1154.773913] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb2835f1-9dc7-488c-8c9c-a8baf9b87692 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.803308] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1154.803308] env[65121]: value = "task-5107375" [ 1154.803308] env[65121]: _type = "Task" [ 1154.803308] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.816774] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107375, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.957973] env[65121]: DEBUG nova.compute.manager [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] [instance: 9f9118f1-c708-469f-839e-b3c4ef52f38f] Received event network-vif-deleted-b3bafe6b-71bc-45d0-bf37-29da60b15de0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1154.958553] env[65121]: DEBUG nova.compute.manager [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1154.958682] env[65121]: DEBUG nova.compute.manager [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing instance network info cache due to event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1154.958922] env[65121]: DEBUG oslo_concurrency.lockutils [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Acquiring lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.959084] env[65121]: DEBUG oslo_concurrency.lockutils [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Acquired lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.959242] env[65121]: DEBUG nova.network.neutron [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1154.964550] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e39bc961-f591-40cd-8f69-74d058bdb277 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "6fe86deb-0903-4769-a05c-b7d0acec0103" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.475s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.007763] env[65121]: DEBUG nova.compute.manager [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Received event network-changed-a2830632-dca1-42e0-be06-922a62ab2307 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1155.007902] env[65121]: DEBUG nova.compute.manager [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Refreshing instance network info cache due to event network-changed-a2830632-dca1-42e0-be06-922a62ab2307. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1155.008215] env[65121]: DEBUG oslo_concurrency.lockutils [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Acquiring lock "refresh_cache-f85ae5c9-ce3f-4ade-b708-4eae79485b20" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.008418] env[65121]: DEBUG oslo_concurrency.lockutils [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Acquired lock "refresh_cache-f85ae5c9-ce3f-4ade-b708-4eae79485b20" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.008631] env[65121]: DEBUG nova.network.neutron [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Refreshing network info cache for port a2830632-dca1-42e0-be06-922a62ab2307 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1155.062291] env[65121]: DEBUG oslo_vmware.api [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107374, 'name': PowerOnVM_Task, 'duration_secs': 0.762912} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.062593] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1155.062796] env[65121]: INFO nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Took 8.12 seconds to spawn the instance on the hypervisor. [ 1155.062985] env[65121]: DEBUG nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1155.064281] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d356570-bf8f-4c33-a2bc-3b85f9dd4ba7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.226249] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a140aa-f976-4cb2-973c-c989a2c85ccb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.240221] env[65121]: DEBUG oslo_vmware.api [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107359, 'name': ReconfigVM_Task, 'duration_secs': 5.86865} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.257530] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.257831] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Reconfigured VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1155.258307] env[65121]: WARNING neutronclient.v2_0.client [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.258699] env[65121]: WARNING neutronclient.v2_0.client [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.259314] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.259762] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.268597] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a93e099-2a9d-42e0-85c1-0b2c099906e8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.274258] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: b22ccce7-c54a-4577-9de0-1fd9c10cd189] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1155.279646] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance '84ae6061-a962-4a74-8661-9718cc4c5346' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1155.314013] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107375, 'name': CreateVM_Task, 'duration_secs': 0.433913} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.314217] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1155.314732] env[65121]: WARNING neutronclient.v2_0.client [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.315115] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.315249] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.315566] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1155.315833] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5da196d7-f536-4be3-aa97-7bf89ef747ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.322472] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1155.322472] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5278ceb1-4fa9-2fa0-a5c8-f283ce96c967" [ 1155.322472] env[65121]: _type = "Task" [ 1155.322472] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.332899] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5278ceb1-4fa9-2fa0-a5c8-f283ce96c967, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.353905] env[65121]: WARNING neutronclient.v2_0.client [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.461624] env[65121]: WARNING neutronclient.v2_0.client [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.462342] env[65121]: WARNING openstack [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.463116] env[65121]: WARNING openstack [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.481674] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1155.481930] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.511856] env[65121]: WARNING neutronclient.v2_0.client [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.512675] env[65121]: WARNING openstack [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.513117] env[65121]: WARNING openstack [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.589509] env[65121]: INFO nova.compute.manager [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Took 20.45 seconds to build instance. [ 1155.772420] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e7d386-f0fb-443e-92ff-822aedaf5b97 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.776837] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: e0d3e3c8-6471-4345-8677-369612674769] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1155.786666] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1155.787129] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8f0cbce-e873-405e-8570-0d1dcac66e9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.789766] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ace2cee-560f-43c5-acb2-caa0d2e4470c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.794523] env[65121]: WARNING openstack [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.794751] env[65121]: WARNING openstack [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.839131] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf98750-5c18-4724-961a-e0679cd7c94f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.842209] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1155.842209] env[65121]: value = "task-5107376" [ 1155.842209] env[65121]: _type = "Task" [ 1155.842209] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.843567] env[65121]: WARNING openstack [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.843914] env[65121]: WARNING openstack [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.861475] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5278ceb1-4fa9-2fa0-a5c8-f283ce96c967, 'name': SearchDatastore_Task, 'duration_secs': 0.010257} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.862283] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.862536] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1155.862763] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.862903] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.863101] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1155.864724] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16dcbb2-5aba-4d89-8cee-56d0fa2bd6d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.872584] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c283b72e-1e8d-4b33-b9f8-2178a0af92b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.874449] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107376, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.886218] env[65121]: DEBUG nova.compute.provider_tree [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.889666] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1155.889846] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1155.890769] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6863da70-ef08-4b35-be86-be7dbc035e91 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.897846] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1155.897846] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]527f3146-df90-3ff8-128c-d53a441ee4cb" [ 1155.897846] env[65121]: _type = "Task" [ 1155.897846] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.910930] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527f3146-df90-3ff8-128c-d53a441ee4cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.986525] env[65121]: DEBUG nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1155.989701] env[65121]: WARNING neutronclient.v2_0.client [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.990366] env[65121]: WARNING openstack [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.990694] env[65121]: WARNING openstack [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.000024] env[65121]: WARNING neutronclient.v2_0.client [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1156.000623] env[65121]: WARNING openstack [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.001026] env[65121]: WARNING openstack [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.090749] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a9ab6605-615f-4b6c-b31e-f70d387e5f54 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.964s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.108770] env[65121]: DEBUG nova.network.neutron [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Updated VIF entry in instance network info cache for port a2830632-dca1-42e0-be06-922a62ab2307. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1156.109140] env[65121]: DEBUG nova.network.neutron [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Updating instance_info_cache with network_info: [{"id": "a2830632-dca1-42e0-be06-922a62ab2307", "address": "fa:16:3e:4f:9c:29", "network": {"id": "cf4d249b-6964-440e-8a07-c9defb8178fa", "bridge": "br-int", "label": "tempest-ServersTestJSON-1496853381-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89110618b6ef4ce68b2ca53e7e168139", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b80dd748-3d7e-4a23-a38d-9e79a3881452", "external-id": "nsx-vlan-transportzone-497", "segmentation_id": 497, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2830632-dc", "ovs_interfaceid": "a2830632-dca1-42e0-be06-922a62ab2307", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1156.140217] env[65121]: DEBUG nova.network.neutron [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updated VIF entry in instance network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1156.140679] env[65121]: DEBUG nova.network.neutron [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91969a04-4a", "ovs_interfaceid": "91969a04-4a8e-4334-829f-3e3ae6828511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1156.281150] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: e2e03e5c-4a71-4555-9cd4-d67e178064ec] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1156.364968] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107376, 'name': PowerOnVM_Task} progress is 86%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.390596] env[65121]: DEBUG nova.scheduler.client.report [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1156.411980] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]527f3146-df90-3ff8-128c-d53a441ee4cb, 'name': SearchDatastore_Task, 'duration_secs': 0.019575} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.413918] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8bba361-c6e5-4627-abc3-8fe3aa989914 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.425184] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1156.425184] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522d95b1-5f75-ed6c-0afd-4f5f032ccd6a" [ 1156.425184] env[65121]: _type = "Task" [ 1156.425184] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.436521] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522d95b1-5f75-ed6c-0afd-4f5f032ccd6a, 'name': SearchDatastore_Task, 'duration_secs': 0.010533} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.436842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1156.437208] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] f85ae5c9-ce3f-4ade-b708-4eae79485b20/f85ae5c9-ce3f-4ade-b708-4eae79485b20.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1156.437516] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8de12105-9f8d-4053-8694-9903763f766d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.448835] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1156.448835] env[65121]: value = "task-5107377" [ 1156.448835] env[65121]: _type = "Task" [ 1156.448835] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.464027] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107377, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.539190] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.612453] env[65121]: DEBUG oslo_concurrency.lockutils [req-11c509f3-db55-4646-bbbf-8d117fc5f5aa req-649846d6-ba7a-423b-8e71-bce671a95ff9 service nova] Releasing lock "refresh_cache-f85ae5c9-ce3f-4ade-b708-4eae79485b20" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1156.643623] env[65121]: DEBUG oslo_concurrency.lockutils [req-89d482ff-373a-4032-888f-d2376f1a7647 req-d9d7340d-e0c9-47f5-b10d-fdff767c1763 service nova] Releasing lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1156.784934] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 3f3bec5b-2834-497c-a454-a152cb992309] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1156.863927] env[65121]: DEBUG oslo_vmware.api [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107376, 'name': PowerOnVM_Task, 'duration_secs': 0.759449} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.864926] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1156.865192] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-337503d4-fa24-4367-bae8-97b61ec2b3f4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance '84ae6061-a962-4a74-8661-9718cc4c5346' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1156.899118] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.899118] env[65121]: DEBUG nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1156.900871] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.633s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.903370] env[65121]: INFO nova.compute.claims [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1156.913307] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.913804] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.914254] env[65121]: DEBUG nova.network.neutron [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1156.960220] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107377, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482872} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.960482] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] f85ae5c9-ce3f-4ade-b708-4eae79485b20/f85ae5c9-ce3f-4ade-b708-4eae79485b20.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1156.960692] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1156.960936] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d86ef25a-371d-4406-9f1f-c2401cf5b0ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.969943] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1156.969943] env[65121]: value = "task-5107378" [ 1156.969943] env[65121]: _type = "Task" [ 1156.969943] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.980506] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107378, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.991695] env[65121]: DEBUG nova.compute.manager [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1156.999708] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.999961] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.290418] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: ab41ae41-e69d-47fb-a31a-16748fd3b0f2] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1157.347488] env[65121]: DEBUG nova.compute.manager [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1157.347626] env[65121]: DEBUG nova.compute.manager [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing instance network info cache due to event network-changed-9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1157.347804] env[65121]: DEBUG oslo_concurrency.lockutils [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Acquiring lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.410090] env[65121]: DEBUG nova.compute.utils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1157.413936] env[65121]: DEBUG nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1157.414052] env[65121]: DEBUG nova.network.neutron [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1157.414413] env[65121]: WARNING neutronclient.v2_0.client [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.414785] env[65121]: WARNING neutronclient.v2_0.client [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.415423] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.415872] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.426517] env[65121]: WARNING neutronclient.v2_0.client [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.427225] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.427313] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.483643] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107378, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.178629} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.485037] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1157.485037] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b944f0-43d5-4e30-b62d-789ecdc166c5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.507825] env[65121]: DEBUG nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1157.520264] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] f85ae5c9-ce3f-4ade-b708-4eae79485b20/f85ae5c9-ce3f-4ade-b708-4eae79485b20.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1157.523208] env[65121]: DEBUG nova.policy [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e20599b4b4f4459eba25b658da1a8484', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a23c876483f34b1db8beec9a669f7822', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1157.528836] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b617a62-1c21-4481-a8dd-ec9c6a969a5e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.545054] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.552860] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1157.552860] env[65121]: value = "task-5107379" [ 1157.552860] env[65121]: _type = "Task" [ 1157.552860] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.565756] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107379, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.636118] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.638469] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.737096] env[65121]: WARNING neutronclient.v2_0.client [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.737830] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.738239] env[65121]: WARNING openstack [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.792481] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: b6f6871d-311c-4adb-824e-2907a12f4224] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1157.859258] env[65121]: INFO nova.network.neutron [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Port f6ae8a69-7565-42c4-96fe-1fbd5e038ebf from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1157.859665] env[65121]: DEBUG nova.network.neutron [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1157.883092] env[65121]: DEBUG nova.network.neutron [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Successfully created port: f714bfb3-87de-4c6b-ba80-8227e6eebf65 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1157.914271] env[65121]: DEBUG nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1158.044273] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.066294] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107379, 'name': ReconfigVM_Task, 'duration_secs': 0.328546} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.066559] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Reconfigured VM instance instance-00000070 to attach disk [datastore1] f85ae5c9-ce3f-4ade-b708-4eae79485b20/f85ae5c9-ce3f-4ade-b708-4eae79485b20.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.067192] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-faf7808e-bd66-4477-83d2-36f225dab2ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.076336] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1158.076336] env[65121]: value = "task-5107380" [ 1158.076336] env[65121]: _type = "Task" [ 1158.076336] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.087857] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107380, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.199636] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ef06c9-29fc-4db0-85c4-9ff19edef3a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.207802] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dff19ae-cb11-4632-9e4e-5c60ae25aea6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.213721] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-8fce40bb-69c1-4fbb-9b0b-c7aece54d179-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.213997] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-8fce40bb-69c1-4fbb-9b0b-c7aece54d179-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1158.214378] env[65121]: DEBUG nova.objects.instance [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'flavor' on Instance uuid 8fce40bb-69c1-4fbb-9b0b-c7aece54d179 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.242074] env[65121]: WARNING neutronclient.v2_0.client [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.242709] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.243060] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.250762] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416a5345-cc0a-4e0c-ac68-54e7165409b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.260792] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4c693f-7bd1-4c31-a3fb-6dfeb6bbf444 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.277535] env[65121]: DEBUG nova.compute.provider_tree [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.296038] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 293f93f2-c01d-42c8-b1a7-3056805c77de] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1158.362375] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1158.364666] env[65121]: DEBUG oslo_concurrency.lockutils [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Acquired lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.364866] env[65121]: DEBUG nova.network.neutron [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Refreshing network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1158.369529] env[65121]: DEBUG nova.objects.instance [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'pci_requests' on Instance uuid 8fce40bb-69c1-4fbb-9b0b-c7aece54d179 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.586589] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107380, 'name': Rename_Task, 'duration_secs': 0.395849} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.586866] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1158.587133] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73221884-db27-4b82-9e02-3b6690c99d2c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.594482] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1158.594482] env[65121]: value = "task-5107381" [ 1158.594482] env[65121]: _type = "Task" [ 1158.594482] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.602949] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107381, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.781580] env[65121]: DEBUG nova.scheduler.client.report [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1158.799556] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 9cf6c29a-3424-4b88-9ba5-8120b124beb6] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1158.867837] env[65121]: DEBUG oslo_concurrency.lockutils [None req-39c9defd-d80d-4712-add8-c46581d4e426 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.182s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.872180] env[65121]: WARNING neutronclient.v2_0.client [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.872849] env[65121]: WARNING openstack [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.873252] env[65121]: WARNING openstack [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.880852] env[65121]: DEBUG nova.objects.base [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Object Instance<8fce40bb-69c1-4fbb-9b0b-c7aece54d179> lazy-loaded attributes: flavor,pci_requests {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1158.880852] env[65121]: DEBUG nova.network.neutron [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1158.881126] env[65121]: WARNING neutronclient.v2_0.client [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.881978] env[65121]: WARNING neutronclient.v2_0.client [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.882114] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.882406] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.926426] env[65121]: DEBUG nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1158.957858] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1158.958119] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1158.958271] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1158.958532] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1158.958688] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1158.958830] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1158.959038] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1158.959194] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1158.959354] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1158.959616] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1158.959794] env[65121]: DEBUG nova.virt.hardware [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1158.960661] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b35980a-d3ec-47c1-b425-13e5d751a080 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.969671] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1090c5d4-7f9b-406b-9bb4-f6d1d5990a91 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.014617] env[65121]: DEBUG nova.policy [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00895ed0ee841b99c242dbaa1562049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35312a302644426f98f127e89a067e75', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1159.106251] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107381, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.286729] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.287289] env[65121]: DEBUG nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1159.290331] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.022s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.294099] env[65121]: INFO nova.compute.claims [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1159.307909] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 7ae12d0f-3f25-42f8-be54-c6e8fb5ded3c] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1159.405210] env[65121]: DEBUG nova.network.neutron [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Successfully updated port: f714bfb3-87de-4c6b-ba80-8227e6eebf65 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1159.424354] env[65121]: WARNING openstack [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.425012] env[65121]: WARNING openstack [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.554418] env[65121]: WARNING neutronclient.v2_0.client [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.555091] env[65121]: WARNING openstack [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.555444] env[65121]: WARNING openstack [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.607049] env[65121]: DEBUG oslo_vmware.api [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107381, 'name': PowerOnVM_Task, 'duration_secs': 0.586653} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.607049] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1159.607049] env[65121]: INFO nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Took 7.87 seconds to spawn the instance on the hypervisor. [ 1159.607049] env[65121]: DEBUG nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1159.610101] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3b33c9-2042-4827-97a0-dd4b2a352fdd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.619044] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.691192] env[65121]: DEBUG nova.network.neutron [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updated VIF entry in instance network info cache for port 9dc8b4e9-1442-43ba-a663-567f89e3a428. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1159.691192] env[65121]: DEBUG nova.network.neutron [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [{"id": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "address": "fa:16:3e:a3:56:9d", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dc8b4e9-14", "ovs_interfaceid": "9dc8b4e9-1442-43ba-a663-567f89e3a428", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.727272] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.727655] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.788133] env[65121]: DEBUG nova.network.neutron [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Port 2acd5dbc-a08c-4ebd-922b-284294369a33 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1159.790036] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.790036] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1159.790036] env[65121]: DEBUG nova.network.neutron [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1159.797030] env[65121]: DEBUG nova.compute.utils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1159.800939] env[65121]: DEBUG nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1159.800939] env[65121]: DEBUG nova.network.neutron [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1159.801235] env[65121]: WARNING neutronclient.v2_0.client [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.801532] env[65121]: WARNING neutronclient.v2_0.client [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.802154] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.804355] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.812052] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 27940143-16b5-4263-b23c-354ed8ea8866] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1159.841762] env[65121]: DEBUG nova.compute.manager [req-dafa7df6-4329-4101-8056-a721b14c6d96 req-6d5bda45-8807-4b7b-92d9-c42ec4bbcd91 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Received event network-vif-plugged-f714bfb3-87de-4c6b-ba80-8227e6eebf65 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1159.842028] env[65121]: DEBUG oslo_concurrency.lockutils [req-dafa7df6-4329-4101-8056-a721b14c6d96 req-6d5bda45-8807-4b7b-92d9-c42ec4bbcd91 service nova] Acquiring lock "3859dd02-9b52-4c71-a41f-c41f47ac1706-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.842936] env[65121]: DEBUG oslo_concurrency.lockutils [req-dafa7df6-4329-4101-8056-a721b14c6d96 req-6d5bda45-8807-4b7b-92d9-c42ec4bbcd91 service nova] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.842936] env[65121]: DEBUG oslo_concurrency.lockutils [req-dafa7df6-4329-4101-8056-a721b14c6d96 req-6d5bda45-8807-4b7b-92d9-c42ec4bbcd91 service nova] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.842936] env[65121]: DEBUG nova.compute.manager [req-dafa7df6-4329-4101-8056-a721b14c6d96 req-6d5bda45-8807-4b7b-92d9-c42ec4bbcd91 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] No waiting events found dispatching network-vif-plugged-f714bfb3-87de-4c6b-ba80-8227e6eebf65 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1159.842936] env[65121]: WARNING nova.compute.manager [req-dafa7df6-4329-4101-8056-a721b14c6d96 req-6d5bda45-8807-4b7b-92d9-c42ec4bbcd91 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Received unexpected event network-vif-plugged-f714bfb3-87de-4c6b-ba80-8227e6eebf65 for instance with vm_state building and task_state spawning. [ 1159.855076] env[65121]: DEBUG nova.policy [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '175d752afbc745258500be8624bfbf14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e491b88dcaad42dcb345b3217bc6f941', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1159.861912] env[65121]: DEBUG nova.compute.manager [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1159.862170] env[65121]: DEBUG nova.compute.manager [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing instance network info cache due to event network-changed-ff1a1ddf-0705-4111-80a1-faa79fe7c947. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1159.862434] env[65121]: DEBUG oslo_concurrency.lockutils [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.862668] env[65121]: DEBUG oslo_concurrency.lockutils [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1159.866645] env[65121]: DEBUG nova.network.neutron [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing network info cache for port ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1159.912332] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.912332] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1159.912332] env[65121]: DEBUG nova.network.neutron [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1160.139521] env[65121]: INFO nova.compute.manager [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Took 21.22 seconds to build instance. [ 1160.153217] env[65121]: DEBUG nova.network.neutron [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Successfully created port: 43502d20-e31f-4bc3-b5ff-4be5428a73a2 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1160.194117] env[65121]: DEBUG oslo_concurrency.lockutils [req-7e7cc691-0955-4ea7-a259-269b47db1095 req-359b586b-da79-4352-88e7-431943ef8f23 service nova] Releasing lock "refresh_cache-3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1160.294591] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.294591] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.294591] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.305018] env[65121]: DEBUG nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1160.315011] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: c2f9fa49-4273-4885-b59d-0635ed739d65] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1160.368558] env[65121]: WARNING neutronclient.v2_0.client [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1160.370331] env[65121]: WARNING openstack [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.370331] env[65121]: WARNING openstack [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.416495] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.417156] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.604449] env[65121]: DEBUG nova.network.neutron [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Successfully updated port: f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1160.631858] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423b5b2a-59fc-410e-bf67-b54bfd5a8a50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.640319] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4be0b505-3448-4d03-96fa-f8572c5e649c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.728s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1160.641511] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c78d07-c13b-4426-abd5-f2011ca7718e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.676483] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3162c2a4-8a8d-45ae-aa08-6ce2168359a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.686349] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87fa58c8-2056-4e4c-9494-acdfcca93248 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.700013] env[65121]: DEBUG nova.compute.provider_tree [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.731386] env[65121]: DEBUG nova.network.neutron [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1160.821995] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: c993d5a4-49d6-43aa-a1f0-4aac91fad953] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1160.867606] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.868032] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.900816] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.901258] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1160.951267] env[65121]: WARNING openstack [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1160.951267] env[65121]: WARNING openstack [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.060863] env[65121]: WARNING neutronclient.v2_0.client [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.061574] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.061913] env[65121]: WARNING openstack [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.087216] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.087933] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.088901] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.108303] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.114344] env[65121]: WARNING neutronclient.v2_0.client [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.115204] env[65121]: WARNING openstack [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.115545] env[65121]: WARNING openstack [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.208141] env[65121]: DEBUG nova.scheduler.client.report [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1161.286580] env[65121]: DEBUG nova.network.neutron [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Updating instance_info_cache with network_info: [{"id": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "address": "fa:16:3e:19:1a:cf", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf714bfb3-87", "ovs_interfaceid": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1161.315063] env[65121]: DEBUG nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1161.322228] env[65121]: DEBUG nova.network.neutron [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1161.324279] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: c4096314-270a-4270-9e1d-5ace8ddbd286] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1161.351107] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1161.351367] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1161.351498] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1161.351669] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1161.351804] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1161.352126] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1161.352359] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1161.352514] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1161.352674] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1161.352826] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1161.352987] env[65121]: DEBUG nova.virt.hardware [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1161.354272] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047e981c-37ec-4eda-9f43-5e8877c0dc36 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.360727] env[65121]: DEBUG nova.network.neutron [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updated VIF entry in instance network info cache for port ff1a1ddf-0705-4111-80a1-faa79fe7c947. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1161.361085] env[65121]: DEBUG nova.network.neutron [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1161.366193] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c058fd8-993d-49c1-b129-f7477dd0b844 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.720394] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.721077] env[65121]: DEBUG nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1161.723480] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.525s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.723703] env[65121]: DEBUG nova.objects.instance [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lazy-loading 'resources' on Instance uuid 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.749373] env[65121]: DEBUG nova.network.neutron [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Successfully updated port: 43502d20-e31f-4bc3-b5ff-4be5428a73a2 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1161.791386] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.791386] env[65121]: DEBUG nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Instance network_info: |[{"id": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "address": "fa:16:3e:19:1a:cf", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf714bfb3-87", "ovs_interfaceid": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1161.791386] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:1a:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f714bfb3-87de-4c6b-ba80-8227e6eebf65', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1161.798547] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating folder: Project (a23c876483f34b1db8beec9a669f7822). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1161.798547] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-54c4cf94-89fa-41f4-9d5d-48a2a79fd769 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.810971] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created folder: Project (a23c876483f34b1db8beec9a669f7822) in parent group-v993268. [ 1161.811199] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating folder: Instances. Parent ref: group-v993575. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1161.811448] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-733fd2ee-10f6-4612-9e68-12e59f5580df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.823843] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created folder: Instances in parent group-v993575. [ 1161.824117] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1161.824312] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1161.824557] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4123e1f-5b9a-4cad-817a-3c366846636c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.840396] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.841872] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: f1a5fdce-a940-43d4-83d7-716786c9ac34] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1161.849886] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1161.849886] env[65121]: value = "task-5107384" [ 1161.849886] env[65121]: _type = "Task" [ 1161.849886] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.862609] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107384, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.866353] env[65121]: DEBUG oslo_concurrency.lockutils [req-18efab91-43aa-441e-bc8f-23bf031fc12d req-3e67c7e9-a3ba-4ee4-ab87-60e04d3566da service nova] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.867102] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1161.867275] env[65121]: DEBUG nova.network.neutron [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1161.935146] env[65121]: DEBUG nova.compute.manager [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Received event network-changed-f714bfb3-87de-4c6b-ba80-8227e6eebf65 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1161.935462] env[65121]: DEBUG nova.compute.manager [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Refreshing instance network info cache due to event network-changed-f714bfb3-87de-4c6b-ba80-8227e6eebf65. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1161.935656] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Acquiring lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.935827] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Acquired lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1161.936090] env[65121]: DEBUG nova.network.neutron [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Refreshing network info cache for port f714bfb3-87de-4c6b-ba80-8227e6eebf65 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1161.956185] env[65121]: DEBUG nova.compute.manager [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-vif-plugged-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1161.956466] env[65121]: DEBUG oslo_concurrency.lockutils [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Acquiring lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.956768] env[65121]: DEBUG oslo_concurrency.lockutils [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.956952] env[65121]: DEBUG oslo_concurrency.lockutils [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.957143] env[65121]: DEBUG nova.compute.manager [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] No waiting events found dispatching network-vif-plugged-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1161.957308] env[65121]: WARNING nova.compute.manager [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received unexpected event network-vif-plugged-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf for instance with vm_state active and task_state None. [ 1161.957461] env[65121]: DEBUG nova.compute.manager [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-changed-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1161.957614] env[65121]: DEBUG nova.compute.manager [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing instance network info cache due to event network-changed-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1161.957766] env[65121]: DEBUG oslo_concurrency.lockutils [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.228028] env[65121]: DEBUG nova.compute.utils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1162.232517] env[65121]: DEBUG nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1162.232517] env[65121]: DEBUG nova.network.neutron [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1162.232517] env[65121]: WARNING neutronclient.v2_0.client [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.232517] env[65121]: WARNING neutronclient.v2_0.client [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.233865] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.234347] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.252794] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "refresh_cache-9475863f-86d5-44e6-ac19-93461ab87743" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.253712] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "refresh_cache-9475863f-86d5-44e6-ac19-93461ab87743" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.253712] env[65121]: DEBUG nova.network.neutron [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1162.291441] env[65121]: DEBUG nova.policy [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e20599b4b4f4459eba25b658da1a8484', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a23c876483f34b1db8beec9a669f7822', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1162.346161] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: cc002961-d742-4255-88c7-f0a5a39424b1] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1162.350142] env[65121]: DEBUG nova.compute.manager [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65121) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1162.374736] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.375186] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.382914] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107384, 'name': CreateVM_Task, 'duration_secs': 0.380602} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.384222] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1162.388061] env[65121]: WARNING neutronclient.v2_0.client [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.388061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.388061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.388316] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1162.389109] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a5d04ab-94f5-4909-8d89-075a719491bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.396489] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1162.396489] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5295befe-0873-b061-e87d-09d98b478517" [ 1162.396489] env[65121]: _type = "Task" [ 1162.396489] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.413052] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5295befe-0873-b061-e87d-09d98b478517, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.437874] env[65121]: WARNING nova.network.neutron [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] 5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd already exists in list: networks containing: ['5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd']. ignoring it [ 1162.440177] env[65121]: WARNING neutronclient.v2_0.client [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.440792] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.441163] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.537749] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.538322] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.579349] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2babeb67-72f8-4782-86de-c60225c640ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.587641] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb8108a-3cde-41a6-8a48-73547b5cf6d3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.632080] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32374a78-1742-4c6c-8474-98a4169c38af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.637795] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.638279] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.650458] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8da6e17-f58c-4d43-bc7b-2d1118a10566 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.668922] env[65121]: DEBUG nova.compute.provider_tree [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.674814] env[65121]: DEBUG nova.network.neutron [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Successfully created port: 10c1c247-c89e-4cd4-9ef3-0bf3621ad21c {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1162.691465] env[65121]: WARNING neutronclient.v2_0.client [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.691878] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.692240] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.732395] env[65121]: DEBUG nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1162.758842] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.759270] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.768601] env[65121]: WARNING neutronclient.v2_0.client [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.769239] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.769608] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.852796] env[65121]: DEBUG nova.network.neutron [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1162.855235] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 5f6e4f46-0745-42c0-a779-6cffb60e21ca] Instance has had 0 of 5 cleanup attempts {{(pid=65121) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1162.863948] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.864186] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.864345] env[65121]: DEBUG nova.compute.manager [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1162.865227] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1c5f78-3091-40cf-80a9-ca4d510860c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.872442] env[65121]: DEBUG nova.compute.manager [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1162.872933] env[65121]: DEBUG nova.objects.instance [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lazy-loading 'flavor' on Instance uuid f85ae5c9-ce3f-4ade-b708-4eae79485b20 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.907315] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5295befe-0873-b061-e87d-09d98b478517, 'name': SearchDatastore_Task, 'duration_secs': 0.014922} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.907614] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1162.907912] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1162.908138] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.908221] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.908395] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1162.908905] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c68b90f1-88ae-4746-a15b-5174405716d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.921041] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.921430] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.932036] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1162.932036] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1162.932036] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fac4720-bb10-4564-96d6-ba2b6cf436a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.936461] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.936836] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.944701] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1162.944701] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52dd1d67-c928-fce6-d071-731f8ab04e11" [ 1162.944701] env[65121]: _type = "Task" [ 1162.944701] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.953357] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dd1d67-c928-fce6-d071-731f8ab04e11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.969816] env[65121]: DEBUG nova.network.neutron [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Updated VIF entry in instance network info cache for port f714bfb3-87de-4c6b-ba80-8227e6eebf65. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1162.970334] env[65121]: DEBUG nova.network.neutron [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Updating instance_info_cache with network_info: [{"id": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "address": "fa:16:3e:19:1a:cf", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf714bfb3-87", "ovs_interfaceid": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.061761] env[65121]: WARNING neutronclient.v2_0.client [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.062377] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1163.062721] env[65121]: WARNING openstack [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1163.085302] env[65121]: WARNING neutronclient.v2_0.client [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.085983] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1163.086358] env[65121]: WARNING openstack [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1163.179028] env[65121]: DEBUG nova.scheduler.client.report [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1163.213735] env[65121]: DEBUG nova.network.neutron [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "address": "fa:16:3e:16:2f:1f", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6ae8a69-75", "ovs_interfaceid": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.244514] env[65121]: DEBUG nova.network.neutron [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Updating instance_info_cache with network_info: [{"id": "43502d20-e31f-4bc3-b5ff-4be5428a73a2", "address": "fa:16:3e:77:c7:44", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43502d20-e3", "ovs_interfaceid": "43502d20-e31f-4bc3-b5ff-4be5428a73a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.359601] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.359601] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Cleaning up deleted instances with incomplete migration {{(pid=65121) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 1163.458653] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52dd1d67-c928-fce6-d071-731f8ab04e11, 'name': SearchDatastore_Task, 'duration_secs': 0.045754} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.459833] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ded720f-9812-4d66-9e2f-0c0ec559d9cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.466804] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1163.466804] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528e20b0-3e3b-cdc8-f47d-1f9f6a863c1c" [ 1163.466804] env[65121]: _type = "Task" [ 1163.466804] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.473660] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Releasing lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.473986] env[65121]: DEBUG nova.compute.manager [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Received event network-vif-plugged-43502d20-e31f-4bc3-b5ff-4be5428a73a2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1163.474260] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Acquiring lock "9475863f-86d5-44e6-ac19-93461ab87743-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.474481] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Lock "9475863f-86d5-44e6-ac19-93461ab87743-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.474642] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Lock "9475863f-86d5-44e6-ac19-93461ab87743-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.474803] env[65121]: DEBUG nova.compute.manager [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] No waiting events found dispatching network-vif-plugged-43502d20-e31f-4bc3-b5ff-4be5428a73a2 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1163.474957] env[65121]: WARNING nova.compute.manager [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Received unexpected event network-vif-plugged-43502d20-e31f-4bc3-b5ff-4be5428a73a2 for instance with vm_state building and task_state spawning. [ 1163.475127] env[65121]: DEBUG nova.compute.manager [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Received event network-changed-43502d20-e31f-4bc3-b5ff-4be5428a73a2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1163.475272] env[65121]: DEBUG nova.compute.manager [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Refreshing instance network info cache due to event network-changed-43502d20-e31f-4bc3-b5ff-4be5428a73a2. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1163.475433] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Acquiring lock "refresh_cache-9475863f-86d5-44e6-ac19-93461ab87743" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.478685] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528e20b0-3e3b-cdc8-f47d-1f9f6a863c1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.495521] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.682052] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.685313] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.958s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.685751] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.687822] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.149s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.689784] env[65121]: INFO nova.compute.claims [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1163.707675] env[65121]: INFO nova.scheduler.client.report [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Deleted allocations for instance 52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731 [ 1163.713633] env[65121]: INFO nova.scheduler.client.report [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted allocations for instance 9f9118f1-c708-469f-839e-b3c4ef52f38f [ 1163.715389] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.716221] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.716391] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1163.719492] env[65121]: DEBUG oslo_concurrency.lockutils [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1163.719978] env[65121]: DEBUG nova.network.neutron [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Refreshing network info cache for port f6ae8a69-7565-42c4-96fe-1fbd5e038ebf {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1163.722101] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574b0cf1-bf76-40ed-a943-02ce1df6472d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.744132] env[65121]: DEBUG nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1163.746836] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1163.747018] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1163.747527] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1163.747762] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1163.747903] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1163.748054] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1163.748400] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.749184] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1163.749184] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1163.749184] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1163.749184] env[65121]: DEBUG nova.virt.hardware [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1163.755780] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Reconfiguring VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1163.757670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "refresh_cache-9475863f-86d5-44e6-ac19-93461ab87743" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.758096] env[65121]: DEBUG nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Instance network_info: |[{"id": "43502d20-e31f-4bc3-b5ff-4be5428a73a2", "address": "fa:16:3e:77:c7:44", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43502d20-e3", "ovs_interfaceid": "43502d20-e31f-4bc3-b5ff-4be5428a73a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1163.758274] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4143f483-0be4-473c-89af-f8a892bafca2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.773967] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Acquired lock "refresh_cache-9475863f-86d5-44e6-ac19-93461ab87743" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1163.774303] env[65121]: DEBUG nova.network.neutron [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Refreshing network info cache for port 43502d20-e31f-4bc3-b5ff-4be5428a73a2 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1163.775858] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:c7:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43502d20-e31f-4bc3-b5ff-4be5428a73a2', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1163.783810] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1163.786766] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1163.788122] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60298db4-b3d9-43f4-89a3-c6472b9d8280 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.806941] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1163.807207] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1163.807332] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1163.807512] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1163.807731] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1163.807902] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1163.808132] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.808286] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1163.808507] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1163.808634] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1163.808806] env[65121]: DEBUG nova.virt.hardware [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1163.809223] env[65121]: DEBUG oslo_vmware.api [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1163.809223] env[65121]: value = "task-5107385" [ 1163.809223] env[65121]: _type = "Task" [ 1163.809223] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.810446] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04abd04-43c2-4311-bffa-bb71fe517c6e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.819219] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1163.819219] env[65121]: value = "task-5107386" [ 1163.819219] env[65121]: _type = "Task" [ 1163.819219] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.829234] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07b6a71-9256-4850-9d8a-c1b42bd56486 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.833850] env[65121]: DEBUG oslo_vmware.api [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107385, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.840992] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107386, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.862392] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.881104] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1163.881352] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f175ab3f-80d0-4d79-889d-cdfc313ede4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.890468] env[65121]: DEBUG oslo_vmware.api [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1163.890468] env[65121]: value = "task-5107387" [ 1163.890468] env[65121]: _type = "Task" [ 1163.890468] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.902052] env[65121]: DEBUG oslo_vmware.api [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.980405] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528e20b0-3e3b-cdc8-f47d-1f9f6a863c1c, 'name': SearchDatastore_Task, 'duration_secs': 0.026809} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.981992] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.981992] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3859dd02-9b52-4c71-a41f-c41f47ac1706.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1163.981992] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17dcb4fa-b62f-4a54-968a-8c14500079a1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.991368] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1163.991368] env[65121]: value = "task-5107388" [ 1163.991368] env[65121]: _type = "Task" [ 1163.991368] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.001665] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.139589] env[65121]: DEBUG nova.compute.manager [req-ea31e160-6854-49df-825f-154de6abd520 req-65a11008-5a48-47b1-aa8a-a311f0a3413b service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Received event network-vif-plugged-10c1c247-c89e-4cd4-9ef3-0bf3621ad21c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1164.139810] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea31e160-6854-49df-825f-154de6abd520 req-65a11008-5a48-47b1-aa8a-a311f0a3413b service nova] Acquiring lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.140026] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea31e160-6854-49df-825f-154de6abd520 req-65a11008-5a48-47b1-aa8a-a311f0a3413b service nova] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.140183] env[65121]: DEBUG oslo_concurrency.lockutils [req-ea31e160-6854-49df-825f-154de6abd520 req-65a11008-5a48-47b1-aa8a-a311f0a3413b service nova] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.140342] env[65121]: DEBUG nova.compute.manager [req-ea31e160-6854-49df-825f-154de6abd520 req-65a11008-5a48-47b1-aa8a-a311f0a3413b service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] No waiting events found dispatching network-vif-plugged-10c1c247-c89e-4cd4-9ef3-0bf3621ad21c {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1164.140494] env[65121]: WARNING nova.compute.manager [req-ea31e160-6854-49df-825f-154de6abd520 req-65a11008-5a48-47b1-aa8a-a311f0a3413b service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Received unexpected event network-vif-plugged-10c1c247-c89e-4cd4-9ef3-0bf3621ad21c for instance with vm_state building and task_state spawning. [ 1164.217373] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c723b4aa-fcba-4e86-a57f-5e42a974e9b0 tempest-ServersNegativeTestJSON-630948171 tempest-ServersNegativeTestJSON-630948171-project-member] Lock "52dfe6ff-2ea4-4ec1-82c5-9e65eafd6731" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.351s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.223640] env[65121]: WARNING neutronclient.v2_0.client [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.224477] env[65121]: WARNING openstack [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.224880] env[65121]: WARNING openstack [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.242442] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b699f7e4-5a8a-4ef2-8331-e79b9e8c1f7f tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9f9118f1-c708-469f-839e-b3c4ef52f38f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.272s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.249861] env[65121]: DEBUG nova.network.neutron [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Successfully updated port: 10c1c247-c89e-4cd4-9ef3-0bf3621ad21c {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1164.276941] env[65121]: WARNING neutronclient.v2_0.client [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.278115] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.278645] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.333302] env[65121]: DEBUG oslo_vmware.api [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107385, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.337062] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107386, 'name': CreateVM_Task, 'duration_secs': 0.381908} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.341882] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1164.343039] env[65121]: WARNING neutronclient.v2_0.client [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.343422] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.343626] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.343876] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1164.344177] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-828abc46-fd85-4ed1-a1ef-2d24c2003dd8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.352712] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1164.352712] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ff2fdd-1945-ebbd-5640-cb29ca16b9d9" [ 1164.352712] env[65121]: _type = "Task" [ 1164.352712] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.361222] env[65121]: WARNING openstack [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.361696] env[65121]: WARNING openstack [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.373945] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ff2fdd-1945-ebbd-5640-cb29ca16b9d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.409275] env[65121]: DEBUG oslo_vmware.api [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107387, 'name': PowerOffVM_Task, 'duration_secs': 0.207669} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.409583] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1164.409784] env[65121]: DEBUG nova.compute.manager [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1164.411919] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be51cc82-9a20-4f77-b8a1-f56bbfcee839 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.420230] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.420677] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.461550] env[65121]: WARNING neutronclient.v2_0.client [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.461550] env[65121]: WARNING openstack [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.461550] env[65121]: WARNING openstack [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.503842] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483335} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.504124] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3859dd02-9b52-4c71-a41f-c41f47ac1706.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1164.504339] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1164.504654] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db1db990-db1e-4ef5-b187-0559a08e1509 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.513301] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1164.513301] env[65121]: value = "task-5107389" [ 1164.513301] env[65121]: _type = "Task" [ 1164.513301] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.524086] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.533642] env[65121]: WARNING neutronclient.v2_0.client [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.534599] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.535044] env[65121]: WARNING openstack [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.574729] env[65121]: DEBUG nova.network.neutron [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updated VIF entry in instance network info cache for port f6ae8a69-7565-42c4-96fe-1fbd5e038ebf. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1164.575295] env[65121]: DEBUG nova.network.neutron [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "address": "fa:16:3e:16:2f:1f", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6ae8a69-75", "ovs_interfaceid": "f6ae8a69-7565-42c4-96fe-1fbd5e038ebf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1164.642344] env[65121]: DEBUG nova.network.neutron [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Updated VIF entry in instance network info cache for port 43502d20-e31f-4bc3-b5ff-4be5428a73a2. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1164.643096] env[65121]: DEBUG nova.network.neutron [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Updating instance_info_cache with network_info: [{"id": "43502d20-e31f-4bc3-b5ff-4be5428a73a2", "address": "fa:16:3e:77:c7:44", "network": {"id": "6ed84af0-b4f0-409d-9256-03fe183ddc15", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-723629382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e491b88dcaad42dcb345b3217bc6f941", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43502d20-e3", "ovs_interfaceid": "43502d20-e31f-4bc3-b5ff-4be5428a73a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1164.752633] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.752840] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.753091] env[65121]: DEBUG nova.network.neutron [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1164.827458] env[65121]: DEBUG oslo_vmware.api [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107385, 'name': ReconfigVM_Task, 'duration_secs': 0.938259} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.827966] env[65121]: WARNING neutronclient.v2_0.client [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.828106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1164.828314] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Reconfigured VM to attach interface {{(pid=65121) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1164.866527] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ff2fdd-1945-ebbd-5640-cb29ca16b9d9, 'name': SearchDatastore_Task, 'duration_secs': 0.057212} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.870511] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1164.870800] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1164.871196] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.871402] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.871558] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1164.872463] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0a2f016-1b0f-4da0-8d69-52f34c96d92b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.886592] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1164.886838] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1164.887829] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb5ec81f-1ce5-4eb7-a532-80957e9f7c29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.896597] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1164.896597] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52269138-b500-4925-7b40-d19d137ce086" [ 1164.896597] env[65121]: _type = "Task" [ 1164.896597] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.909568] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52269138-b500-4925-7b40-d19d137ce086, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.937551] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9f6770bb-04ef-4e5f-bf79-1d1c8590017c tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.073s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.997191] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f682967e-da80-4031-84e2-93504abd97ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.006118] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236a7376-0867-4954-b71b-9dff20bbf700 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.044689] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79b750a-6901-455d-a3e2-8dc12904f15b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.053548] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076692} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.056062] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1165.057012] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6667f51f-a9a3-4cb3-b182-4aaa55ca322e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.061362] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248e6187-51aa-48db-8051-edd41bc07077 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.076014] env[65121]: DEBUG nova.compute.provider_tree [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.086858] env[65121]: DEBUG oslo_concurrency.lockutils [req-1d89e35c-86cd-47f2-9f71-76a8ea9a9c1c req-0e8c2f5b-c92d-4bc1-a596-b41bb9847bca service nova] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.096011] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3859dd02-9b52-4c71-a41f-c41f47ac1706.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1165.097213] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a5a3d1e-651e-4802-a1e4-408d7dd11617 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.118198] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1165.118198] env[65121]: value = "task-5107390" [ 1165.118198] env[65121]: _type = "Task" [ 1165.118198] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.127751] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107390, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.146037] env[65121]: DEBUG oslo_concurrency.lockutils [req-64f4065e-b3c3-4023-8ccb-f69aab13a51d req-d3e7fe02-07a2-4359-9da5-00d658cf3744 service nova] Releasing lock "refresh_cache-9475863f-86d5-44e6-ac19-93461ab87743" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.256481] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.256876] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.296725] env[65121]: DEBUG nova.network.neutron [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1165.333047] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1565cd8e-335b-4170-90fc-be4d7b975ca8 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-8fce40bb-69c1-4fbb-9b0b-c7aece54d179-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.119s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1165.338325] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.338325] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.415289] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52269138-b500-4925-7b40-d19d137ce086, 'name': SearchDatastore_Task, 'duration_secs': 0.026743} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.417839] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5d09a2f-e49b-4827-a916-5017754a155c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.423055] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1165.423055] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ba9df4-5c4e-0990-5aed-43de5149f5ac" [ 1165.423055] env[65121]: _type = "Task" [ 1165.423055] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.432782] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ba9df4-5c4e-0990-5aed-43de5149f5ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.445556] env[65121]: WARNING neutronclient.v2_0.client [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1165.446552] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.446936] env[65121]: WARNING openstack [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.570980] env[65121]: DEBUG nova.network.neutron [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Updating instance_info_cache with network_info: [{"id": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "address": "fa:16:3e:eb:2a:3d", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10c1c247-c8", "ovs_interfaceid": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1165.597630] env[65121]: DEBUG nova.scheduler.client.report [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1165.631042] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107390, 'name': ReconfigVM_Task, 'duration_secs': 0.277822} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.631313] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3859dd02-9b52-4c71-a41f-c41f47ac1706.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.631940] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80a97dfe-c8c6-4101-9d7f-805b17817a4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.638515] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1165.638515] env[65121]: value = "task-5107391" [ 1165.638515] env[65121]: _type = "Task" [ 1165.638515] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.647879] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107391, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.934688] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ba9df4-5c4e-0990-5aed-43de5149f5ac, 'name': SearchDatastore_Task, 'duration_secs': 0.017507} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.935503] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.935503] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 9475863f-86d5-44e6-ac19-93461ab87743/9475863f-86d5-44e6-ac19-93461ab87743.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1165.935503] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac6ec6eb-cb66-4c99-a1a4-2a48a63e2ebf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.942775] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1165.942775] env[65121]: value = "task-5107392" [ 1165.942775] env[65121]: _type = "Task" [ 1165.942775] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.952143] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.075409] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.075795] env[65121]: DEBUG nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Instance network_info: |[{"id": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "address": "fa:16:3e:eb:2a:3d", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10c1c247-c8", "ovs_interfaceid": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1166.076293] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:2a:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10c1c247-c89e-4cd4-9ef3-0bf3621ad21c', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1166.084278] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1166.084543] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1166.084775] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ce011a5-0927-42a3-90b2-b8eee29fbde1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.100278] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.100754] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.100962] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.101152] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.101312] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.104357] env[65121]: INFO nova.compute.manager [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Terminating instance [ 1166.106144] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.106663] env[65121]: DEBUG nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1166.117011] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.572s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.119233] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1166.119233] env[65121]: value = "task-5107393" [ 1166.119233] env[65121]: _type = "Task" [ 1166.119233] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.133881] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107393, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.152151] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107391, 'name': Rename_Task, 'duration_secs': 0.143538} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.152655] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1166.153046] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f192d858-ca27-4650-9cfd-c30d685e5763 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.161480] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1166.161480] env[65121]: value = "task-5107394" [ 1166.161480] env[65121]: _type = "Task" [ 1166.161480] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.171995] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.181288] env[65121]: DEBUG nova.compute.manager [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Received event network-changed-10c1c247-c89e-4cd4-9ef3-0bf3621ad21c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1166.182149] env[65121]: DEBUG nova.compute.manager [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Refreshing instance network info cache due to event network-changed-10c1c247-c89e-4cd4-9ef3-0bf3621ad21c. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1166.182149] env[65121]: DEBUG oslo_concurrency.lockutils [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Acquiring lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.182149] env[65121]: DEBUG oslo_concurrency.lockutils [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Acquired lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.182149] env[65121]: DEBUG nova.network.neutron [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Refreshing network info cache for port 10c1c247-c89e-4cd4-9ef3-0bf3621ad21c {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1166.456984] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107392, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.613733] env[65121]: DEBUG nova.compute.utils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1166.615812] env[65121]: DEBUG nova.compute.manager [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1166.615929] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.616244] env[65121]: DEBUG nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1166.616459] env[65121]: DEBUG nova.network.neutron [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1166.616896] env[65121]: WARNING neutronclient.v2_0.client [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.617134] env[65121]: WARNING neutronclient.v2_0.client [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.621322] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.621692] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.629672] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27411741-9492-461a-8356-fea75e693771 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.634982] env[65121]: INFO nova.compute.claims [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1166.648545] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107393, 'name': CreateVM_Task, 'duration_secs': 0.469061} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.651657] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1166.651657] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1166.651946] env[65121]: WARNING neutronclient.v2_0.client [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.652198] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.652345] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.652787] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1166.653524] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29a9ca33-e373-4e6a-9e0c-5ce2fcac5754 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.654987] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bff08e3-fde6-4c5e-88f6-4d14df77994b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.660949] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1166.660949] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c0d66d-7fba-1daa-f3ed-55f7e2f4d7dc" [ 1166.660949] env[65121]: _type = "Task" [ 1166.660949] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.680064] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c0d66d-7fba-1daa-f3ed-55f7e2f4d7dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.680341] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107394, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.684187] env[65121]: DEBUG nova.policy [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4853c778ab79419da0e9de23ee3617a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f056059180af48bba4587006efb0b211', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1166.687700] env[65121]: WARNING neutronclient.v2_0.client [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.688580] env[65121]: WARNING openstack [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.689158] env[65121]: WARNING openstack [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.732390] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.732614] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.732790] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleting the datastore file [datastore1] f85ae5c9-ce3f-4ade-b708-4eae79485b20 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.733105] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc61de5f-c7e7-49fa-8c74-ad8497b4d5bd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.741250] env[65121]: DEBUG oslo_vmware.api [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1166.741250] env[65121]: value = "task-5107396" [ 1166.741250] env[65121]: _type = "Task" [ 1166.741250] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.752714] env[65121]: DEBUG oslo_vmware.api [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107396, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.811130] env[65121]: WARNING openstack [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.811564] env[65121]: WARNING openstack [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.927397] env[65121]: WARNING neutronclient.v2_0.client [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.928062] env[65121]: WARNING openstack [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.928403] env[65121]: WARNING openstack [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.956125] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.6581} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.956414] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 9475863f-86d5-44e6-ac19-93461ab87743/9475863f-86d5-44e6-ac19-93461ab87743.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1166.956649] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1166.956913] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-241bf4f0-d4ef-4e73-b760-10cd952d74d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.965394] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1166.965394] env[65121]: value = "task-5107397" [ 1166.965394] env[65121]: _type = "Task" [ 1166.965394] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.988229] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107397, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.032880] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "interface-8fce40bb-69c1-4fbb-9b0b-c7aece54d179-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.033262] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-8fce40bb-69c1-4fbb-9b0b-c7aece54d179-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.042923] env[65121]: DEBUG nova.network.neutron [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Successfully created port: 3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1167.077941] env[65121]: DEBUG nova.network.neutron [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Updated VIF entry in instance network info cache for port 10c1c247-c89e-4cd4-9ef3-0bf3621ad21c. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1167.078425] env[65121]: DEBUG nova.network.neutron [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Updating instance_info_cache with network_info: [{"id": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "address": "fa:16:3e:eb:2a:3d", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10c1c247-c8", "ovs_interfaceid": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1167.131983] env[65121]: DEBUG nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1167.146058] env[65121]: INFO nova.compute.resource_tracker [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating resource usage from migration 94b4d03a-9745-4cc1-bed1-3ff1d03f3bfb [ 1167.190788] env[65121]: DEBUG oslo_vmware.api [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107394, 'name': PowerOnVM_Task, 'duration_secs': 0.786667} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.191100] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c0d66d-7fba-1daa-f3ed-55f7e2f4d7dc, 'name': SearchDatastore_Task, 'duration_secs': 0.017758} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.194042] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1167.194215] env[65121]: INFO nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Took 8.27 seconds to spawn the instance on the hypervisor. [ 1167.194389] env[65121]: DEBUG nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1167.194748] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.194957] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1167.195204] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.195342] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1167.195514] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1167.196951] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2fbe62-a227-45c5-8b7f-7e86112b9253 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.199586] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0dc864cb-4fbd-464b-8bf7-b2c30c8fa527 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.214452] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1167.214920] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1167.216651] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52a995d1-78af-499a-80a2-83e02f28f385 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.223025] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1167.223025] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5280549d-23bb-9886-8d26-4ce5b7b91ec2" [ 1167.223025] env[65121]: _type = "Task" [ 1167.223025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.233892] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5280549d-23bb-9886-8d26-4ce5b7b91ec2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.253349] env[65121]: DEBUG oslo_vmware.api [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107396, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156098} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.253611] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.253787] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1167.253954] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1167.254148] env[65121]: INFO nova.compute.manager [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1167.254386] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1167.254620] env[65121]: DEBUG nova.compute.manager [-] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1167.254726] env[65121]: DEBUG nova.network.neutron [-] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1167.254979] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1167.255530] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1167.255782] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1167.296448] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1167.443030] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c55244-d5dd-472c-8150-82e3916bd208 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.452384] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f674c663-0467-4cb5-b2ae-7e762e38a411 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.489829] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a3f704-507f-49c1-8be9-a07ec874d9c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.498805] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107397, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.250155} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.501227] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1167.502062] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b03567-be04-41ca-bbab-bedef311040b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.506589] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8fc647-9967-4031-9607-8738af335285 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.533747] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 9475863f-86d5-44e6-ac19-93461ab87743/9475863f-86d5-44e6-ac19-93461ab87743.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.545337] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a1397dc-32e1-41e0-8b13-1e63a52ca58d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.562944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.563119] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1167.563652] env[65121]: DEBUG nova.compute.provider_tree [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.569024] env[65121]: DEBUG nova.network.neutron [-] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1167.569024] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f6c81b-4980-4db7-b906-009f87272615 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.589501] env[65121]: DEBUG oslo_concurrency.lockutils [req-a831b358-5de2-49c7-9f4d-25ef22513e10 req-81507ce9-b1f8-415f-b20c-6d33989e46aa service nova] Releasing lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.591727] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43ea656-8988-4381-b2a8-c162461d4cab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.595071] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1167.595071] env[65121]: value = "task-5107398" [ 1167.595071] env[65121]: _type = "Task" [ 1167.595071] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.615658] env[65121]: WARNING neutronclient.v2_0.client [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1167.621365] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Reconfiguring VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1167.622591] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b18b095-9262-47ac-bfea-2cec6028468b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.644131] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.655028] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1167.655028] env[65121]: value = "task-5107399" [ 1167.655028] env[65121]: _type = "Task" [ 1167.655028] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.662020] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.731056] env[65121]: INFO nova.compute.manager [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Took 17.54 seconds to build instance. [ 1167.741616] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5280549d-23bb-9886-8d26-4ce5b7b91ec2, 'name': SearchDatastore_Task, 'duration_secs': 0.011366} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.743393] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-491f15e5-a130-496c-babb-c6093ffc15b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.751067] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1167.751067] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ae81ce-12e5-2a52-43f8-3df32e15f41a" [ 1167.751067] env[65121]: _type = "Task" [ 1167.751067] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.764558] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ae81ce-12e5-2a52-43f8-3df32e15f41a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.072849] env[65121]: DEBUG nova.scheduler.client.report [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1168.076592] env[65121]: INFO nova.compute.manager [-] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Took 0.82 seconds to deallocate network for instance. [ 1168.106201] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.147329] env[65121]: DEBUG nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1168.171430] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.181763] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1168.181763] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1168.181763] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1168.181763] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1168.182083] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1168.182083] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1168.182236] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1168.182385] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1168.182546] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1168.182735] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1168.182851] env[65121]: DEBUG nova.virt.hardware [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1168.183841] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2f8f02-d22e-497a-a373-d71fabab1459 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.196603] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8c7786-795e-49a5-9e2e-334151f4db4e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.233643] env[65121]: DEBUG oslo_concurrency.lockutils [None req-fb0b9a3b-cdbd-459b-89e1-655481bb281f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.058s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.242035] env[65121]: DEBUG nova.compute.manager [req-c8ad845e-feb3-468c-a4ff-8a02e3ff896a req-092eabd4-5b66-48d8-a728-a7e0378b22ba service nova] [instance: f85ae5c9-ce3f-4ade-b708-4eae79485b20] Received event network-vif-deleted-a2830632-dca1-42e0-be06-922a62ab2307 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1168.265542] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ae81ce-12e5-2a52-43f8-3df32e15f41a, 'name': SearchDatastore_Task, 'duration_secs': 0.012294} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.266021] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1168.266163] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/d49a301c-c0d5-476e-9f1f-b8f98cf6adb7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1168.266345] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-adbf73b9-cb78-4e9c-a12a-44998bfb4329 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.275475] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1168.275475] env[65121]: value = "task-5107400" [ 1168.275475] env[65121]: _type = "Task" [ 1168.275475] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.287187] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107400, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.548496] env[65121]: DEBUG nova.compute.manager [req-4d536562-b2f1-4048-87bd-47254c6f44df req-76ba7e69-6f65-4fac-978b-a2f682f52ba3 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Received event network-vif-plugged-3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1168.549166] env[65121]: DEBUG oslo_concurrency.lockutils [req-4d536562-b2f1-4048-87bd-47254c6f44df req-76ba7e69-6f65-4fac-978b-a2f682f52ba3 service nova] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.549379] env[65121]: DEBUG oslo_concurrency.lockutils [req-4d536562-b2f1-4048-87bd-47254c6f44df req-76ba7e69-6f65-4fac-978b-a2f682f52ba3 service nova] Lock "138e48fe-0221-4173-801b-ee53b2bb98df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.549565] env[65121]: DEBUG oslo_concurrency.lockutils [req-4d536562-b2f1-4048-87bd-47254c6f44df req-76ba7e69-6f65-4fac-978b-a2f682f52ba3 service nova] Lock "138e48fe-0221-4173-801b-ee53b2bb98df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.549776] env[65121]: DEBUG nova.compute.manager [req-4d536562-b2f1-4048-87bd-47254c6f44df req-76ba7e69-6f65-4fac-978b-a2f682f52ba3 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] No waiting events found dispatching network-vif-plugged-3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1168.549942] env[65121]: WARNING nova.compute.manager [req-4d536562-b2f1-4048-87bd-47254c6f44df req-76ba7e69-6f65-4fac-978b-a2f682f52ba3 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Received unexpected event network-vif-plugged-3987861c-5f58-4f3f-868e-31452ab11d4d for instance with vm_state building and task_state spawning. [ 1168.577535] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.460s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.577828] env[65121]: INFO nova.compute.manager [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Migrating [ 1168.590561] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.547s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.592348] env[65121]: INFO nova.compute.claims [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1168.596844] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.616054] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.664493] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.692356] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.692537] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.714280] env[65121]: DEBUG nova.network.neutron [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Successfully updated port: 3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1168.790979] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107400, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503762} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.791312] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/d49a301c-c0d5-476e-9f1f-b8f98cf6adb7.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1168.791522] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1168.791806] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd9b17d0-af43-4abd-ab4f-6144a18041de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.799461] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1168.799461] env[65121]: value = "task-5107401" [ 1168.799461] env[65121]: _type = "Task" [ 1168.799461] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.810599] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107401, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.106344] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.106486] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.106571] env[65121]: DEBUG nova.network.neutron [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1169.119953] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107398, 'name': ReconfigVM_Task, 'duration_secs': 1.145825} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.120296] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 9475863f-86d5-44e6-ac19-93461ab87743/9475863f-86d5-44e6-ac19-93461ab87743.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1169.121235] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-86812451-dbd0-42f7-8c3e-e3dbd1214366 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.129945] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1169.129945] env[65121]: value = "task-5107402" [ 1169.129945] env[65121]: _type = "Task" [ 1169.129945] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.144089] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107402, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.165967] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.197078] env[65121]: INFO nova.compute.manager [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Detaching volume 41abb581-3841-4c0e-b79e-9820e1b241f1 [ 1169.219268] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.219268] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.219268] env[65121]: DEBUG nova.network.neutron [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1169.243411] env[65121]: INFO nova.virt.block_device [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Attempting to driver detach volume 41abb581-3841-4c0e-b79e-9820e1b241f1 from mountpoint /dev/sdb [ 1169.244097] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1169.244097] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1169.244888] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dab9b9d-66ea-4fb2-a3e0-2c9aa162ae02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.271810] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1097226e-182b-4c41-944f-17026f804ff0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.280684] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b0a422-3afd-4b30-a816-b4ee0ce8cb7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.307996] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee64c85-6699-410b-b91d-7a720ddb4ffa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.317081] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107401, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.356655} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.330622] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1169.331101] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The volume has not been displaced from its original location: [datastore2] volume-41abb581-3841-4c0e-b79e-9820e1b241f1/volume-41abb581-3841-4c0e-b79e-9820e1b241f1.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1169.337594] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1169.338754] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762b8163-1b6e-46e3-a7d5-8211937a5de2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.341678] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5aef1a6b-c6cf-4aa6-b83f-5a95e48c829c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.377890] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/d49a301c-c0d5-476e-9f1f-b8f98cf6adb7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1169.379720] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1020caf4-420a-46d8-b782-37110283b0c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.398025] env[65121]: DEBUG oslo_vmware.api [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1169.398025] env[65121]: value = "task-5107403" [ 1169.398025] env[65121]: _type = "Task" [ 1169.398025] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.402991] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1169.402991] env[65121]: value = "task-5107404" [ 1169.402991] env[65121]: _type = "Task" [ 1169.402991] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.412613] env[65121]: DEBUG oslo_vmware.api [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.419087] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107404, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.613156] env[65121]: WARNING neutronclient.v2_0.client [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1169.614052] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.614588] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.646728] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107402, 'name': Rename_Task, 'duration_secs': 0.345017} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.646728] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1169.646728] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66ea8101-9f5a-4a03-b780-6b2e8f557301 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.656430] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1169.656430] env[65121]: value = "task-5107405" [ 1169.656430] env[65121]: _type = "Task" [ 1169.656430] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.679073] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107405, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.679073] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.724637] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.725081] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.811783] env[65121]: DEBUG nova.network.neutron [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1169.831192] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.831683] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.847917] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.848383] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.917493] env[65121]: DEBUG oslo_vmware.api [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107403, 'name': ReconfigVM_Task, 'duration_secs': 0.292514} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.921353] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1169.926633] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107404, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.929864] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9c9cd3f-73ff-4ebf-9686-32c4e57e99be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.942735] env[65121]: WARNING neutronclient.v2_0.client [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1169.943510] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.943852] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.955329] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d62bea8-6fd5-4ee3-810b-c9c56b8edfe6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.960980] env[65121]: DEBUG oslo_vmware.api [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1169.960980] env[65121]: value = "task-5107406" [ 1169.960980] env[65121]: _type = "Task" [ 1169.960980] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.967548] env[65121]: WARNING neutronclient.v2_0.client [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1169.968700] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.969352] env[65121]: WARNING openstack [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.982166] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88dc239-1ea3-489c-bad3-a311eefec33d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.990136] env[65121]: DEBUG oslo_vmware.api [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107406, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.027737] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9a482a-b51b-4262-8313-b6808b6ea7f3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.038161] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d345a7-7d2d-4daf-a98b-62a7bfa830da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.054974] env[65121]: DEBUG nova.compute.provider_tree [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.096757] env[65121]: DEBUG nova.network.neutron [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance_info_cache with network_info: [{"id": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "address": "fa:16:3e:22:e8:da", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd86b7885-6e", "ovs_interfaceid": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.107377] env[65121]: DEBUG nova.network.neutron [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Updating instance_info_cache with network_info: [{"id": "3987861c-5f58-4f3f-868e-31452ab11d4d", "address": "fa:16:3e:75:a0:1a", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3987861c-5f", "ovs_interfaceid": "3987861c-5f58-4f3f-868e-31452ab11d4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.174655] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.174993] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107405, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.419921] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107404, 'name': ReconfigVM_Task, 'duration_secs': 0.750635} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.420268] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Reconfigured VM instance instance-00000073 to attach disk [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/d49a301c-c0d5-476e-9f1f-b8f98cf6adb7.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1170.420990] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9c6cecc-c365-4c88-b68d-cca7d38be91d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.429552] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1170.429552] env[65121]: value = "task-5107407" [ 1170.429552] env[65121]: _type = "Task" [ 1170.429552] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.440359] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107407, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.472065] env[65121]: DEBUG oslo_vmware.api [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107406, 'name': ReconfigVM_Task, 'duration_secs': 0.165038} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.472065] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993555', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'name': 'volume-41abb581-3841-4c0e-b79e-9820e1b241f1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1fbd3a12-82d3-458d-a590-0f5c0c70ae55', 'attached_at': '', 'detached_at': '', 'volume_id': '41abb581-3841-4c0e-b79e-9820e1b241f1', 'serial': '41abb581-3841-4c0e-b79e-9820e1b241f1'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1170.566922] env[65121]: DEBUG nova.scheduler.client.report [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1170.582998] env[65121]: DEBUG nova.compute.manager [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Received event network-changed-3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1170.583249] env[65121]: DEBUG nova.compute.manager [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Refreshing instance network info cache due to event network-changed-3987861c-5f58-4f3f-868e-31452ab11d4d. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1170.583409] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Acquiring lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.600181] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1170.610126] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1170.610596] env[65121]: DEBUG nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Instance network_info: |[{"id": "3987861c-5f58-4f3f-868e-31452ab11d4d", "address": "fa:16:3e:75:a0:1a", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3987861c-5f", "ovs_interfaceid": "3987861c-5f58-4f3f-868e-31452ab11d4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1170.611023] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Acquired lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1170.611250] env[65121]: DEBUG nova.network.neutron [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Refreshing network info cache for port 3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1170.612940] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:a0:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3987861c-5f58-4f3f-868e-31452ab11d4d', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1170.621210] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1170.623035] env[65121]: WARNING neutronclient.v2_0.client [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.624023] env[65121]: WARNING openstack [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.624502] env[65121]: WARNING openstack [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.633113] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1170.633983] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4d48346-9f89-44b4-a7a7-1e0933ed1b09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.657869] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1170.657869] env[65121]: value = "task-5107408" [ 1170.657869] env[65121]: _type = "Task" [ 1170.657869] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.675047] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.681084] env[65121]: DEBUG oslo_vmware.api [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107405, 'name': PowerOnVM_Task, 'duration_secs': 0.838109} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.681345] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107408, 'name': CreateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.683753] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1170.683970] env[65121]: INFO nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Took 9.37 seconds to spawn the instance on the hypervisor. [ 1170.684163] env[65121]: DEBUG nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1170.685429] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5baa6044-e688-45ee-a964-c504c68e4984 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.768922] env[65121]: WARNING openstack [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.769545] env[65121]: WARNING openstack [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.857060] env[65121]: WARNING neutronclient.v2_0.client [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.857845] env[65121]: WARNING openstack [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.858204] env[65121]: WARNING openstack [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.940178] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107407, 'name': Rename_Task, 'duration_secs': 0.216857} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.942368] env[65121]: DEBUG nova.network.neutron [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Updated VIF entry in instance network info cache for port 3987861c-5f58-4f3f-868e-31452ab11d4d. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1170.942723] env[65121]: DEBUG nova.network.neutron [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Updating instance_info_cache with network_info: [{"id": "3987861c-5f58-4f3f-868e-31452ab11d4d", "address": "fa:16:3e:75:a0:1a", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3987861c-5f", "ovs_interfaceid": "3987861c-5f58-4f3f-868e-31452ab11d4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.943944] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1170.944505] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51a37d96-39bc-482e-ba85-4b8eb73586d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.952753] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1170.952753] env[65121]: value = "task-5107409" [ 1170.952753] env[65121]: _type = "Task" [ 1170.952753] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.963074] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107409, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.021071] env[65121]: DEBUG nova.objects.instance [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'flavor' on Instance uuid 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.072621] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.073386] env[65121]: DEBUG nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1171.077289] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 7.582s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.176541] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107408, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.176815] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.206011] env[65121]: INFO nova.compute.manager [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Took 19.99 seconds to build instance. [ 1171.445626] env[65121]: DEBUG oslo_concurrency.lockutils [req-7f58f85b-6c9a-46f4-8e7e-53943d12f0cf req-a58833b1-1b83-4037-9f88-5f68e5b3d922 service nova] Releasing lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1171.465645] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107409, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.581313] env[65121]: DEBUG nova.compute.utils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1171.586043] env[65121]: DEBUG nova.objects.instance [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'migration_context' on Instance uuid 84ae6061-a962-4a74-8661-9718cc4c5346 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.586043] env[65121]: DEBUG nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1171.586043] env[65121]: DEBUG nova.network.neutron [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1171.586043] env[65121]: WARNING neutronclient.v2_0.client [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.586043] env[65121]: WARNING neutronclient.v2_0.client [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.586043] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1171.586424] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1171.667749] env[65121]: DEBUG nova.policy [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7f1d5cb6ec4c0f8a13f836eb712340', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ce7f6698e214d73ae43427601058af8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1171.681983] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.685739] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107408, 'name': CreateVM_Task, 'duration_secs': 0.551003} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.685739] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1171.685950] env[65121]: WARNING neutronclient.v2_0.client [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.686417] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.686417] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1171.686769] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1171.687074] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1dffa58-1050-4f1a-8aec-9a05716e06e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.693346] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1171.693346] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5236b9fe-45b2-387e-17f7-e92454af1b44" [ 1171.693346] env[65121]: _type = "Task" [ 1171.693346] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.704134] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5236b9fe-45b2-387e-17f7-e92454af1b44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.708847] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0985fd10-be4d-4a47-99de-185903070f9b tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9475863f-86d5-44e6-ac19-93461ab87743" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.938s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.966300] env[65121]: DEBUG oslo_vmware.api [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107409, 'name': PowerOnVM_Task, 'duration_secs': 0.706809} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.966300] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1171.966300] env[65121]: INFO nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Took 8.22 seconds to spawn the instance on the hypervisor. [ 1171.966494] env[65121]: DEBUG nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1171.967413] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b86fb5a-0a3f-4020-a7fb-bbf84329cd29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.021060] env[65121]: DEBUG nova.network.neutron [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Successfully created port: 2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1172.028919] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0a1b12b1-1a62-4fe7-9472-675d839d44c7 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.336s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.085210] env[65121]: DEBUG nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1172.120161] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5937cda0-d390-4892-8b26-95905fffe42b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.142795] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance '9a248b47-a34b-4a98-baf0-aa971b5aca94' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1172.175403] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.208866] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5236b9fe-45b2-387e-17f7-e92454af1b44, 'name': SearchDatastore_Task, 'duration_secs': 0.025405} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.212848] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1172.214354] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1172.214684] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.214840] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1172.215063] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1172.216277] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6868f763-72b4-4209-86e8-9bbaae549347 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.233517] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1172.233815] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1172.241030] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19deab04-c37b-4e43-b98a-54ef2201a80a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.247830] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1172.247830] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5290fd54-adfd-6e39-3c21-e7bb8f4a8ca5" [ 1172.247830] env[65121]: _type = "Task" [ 1172.247830] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.258868] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5290fd54-adfd-6e39-3c21-e7bb8f4a8ca5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.460367] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b81b87-6983-4f3f-b5eb-e0623c3233b6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.469563] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d2e14d-b76d-42ac-84ea-ec9af0603e47 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.511720] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d499f6-8ed2-45ec-a86a-dae282626658 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.514323] env[65121]: INFO nova.compute.manager [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Took 21.29 seconds to build instance. [ 1172.521522] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a22622-1a63-4b0c-8c10-89360ba8dada {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.538455] env[65121]: DEBUG nova.compute.provider_tree [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.651898] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.652240] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93773538-c662-4e55-a89a-1098a1d77c52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.662031] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1172.662031] env[65121]: value = "task-5107410" [ 1172.662031] env[65121]: _type = "Task" [ 1172.662031] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.677600] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107410, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.681493] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.763026] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5290fd54-adfd-6e39-3c21-e7bb8f4a8ca5, 'name': SearchDatastore_Task, 'duration_secs': 0.026801} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.764264] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4341492d-8364-4f43-ae87-aeb73ad3abfe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.772706] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1172.772706] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fc20d5-5fb0-cfc2-21d3-85275a84da64" [ 1172.772706] env[65121]: _type = "Task" [ 1172.772706] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.782928] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fc20d5-5fb0-cfc2-21d3-85275a84da64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.017795] env[65121]: DEBUG oslo_concurrency.lockutils [None req-76a497bd-d599-49e3-b0f3-1630cbe6bc4c tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.220s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.044034] env[65121]: DEBUG nova.scheduler.client.report [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1173.100953] env[65121]: DEBUG nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1173.127431] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1173.127722] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1173.127842] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1173.128072] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1173.128279] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1173.128436] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1173.128713] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1173.128881] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1173.129064] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1173.129221] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1173.129382] env[65121]: DEBUG nova.virt.hardware [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1173.130743] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17d65ae-ed08-4af5-a6d3-8c337b4f79c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.140044] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed681109-60d4-426e-85e5-964d1e172a05 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.172952] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.179286] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107410, 'name': PowerOffVM_Task, 'duration_secs': 0.226157} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.179633] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1173.179878] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance '9a248b47-a34b-4a98-baf0-aa971b5aca94' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1173.187307] env[65121]: INFO nova.compute.manager [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Rescuing [ 1173.187519] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.187650] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1173.187833] env[65121]: DEBUG nova.network.neutron [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1173.284520] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52fc20d5-5fb0-cfc2-21d3-85275a84da64, 'name': SearchDatastore_Task, 'duration_secs': 0.012791} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.284933] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1173.285228] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 138e48fe-0221-4173-801b-ee53b2bb98df/138e48fe-0221-4173-801b-ee53b2bb98df.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1173.285501] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2afca544-ab77-4681-9504-004f2c6a8679 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.294558] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1173.294558] env[65121]: value = "task-5107411" [ 1173.294558] env[65121]: _type = "Task" [ 1173.294558] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.306045] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.369020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9475863f-86d5-44e6-ac19-93461ab87743" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.369677] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9475863f-86d5-44e6-ac19-93461ab87743" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.370181] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "9475863f-86d5-44e6-ac19-93461ab87743-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.370571] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9475863f-86d5-44e6-ac19-93461ab87743-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.370893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9475863f-86d5-44e6-ac19-93461ab87743-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.374722] env[65121]: INFO nova.compute.manager [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Terminating instance [ 1173.456926] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.461087] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.461497] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.461776] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.462018] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.465810] env[65121]: INFO nova.compute.manager [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Terminating instance [ 1173.539604] env[65121]: DEBUG nova.compute.manager [req-66b1321f-dac2-41a0-9fcf-2f0867487e63 req-b830e666-9fc3-4e93-8495-aa2dd8dabc44 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Received event network-vif-plugged-2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1173.539967] env[65121]: DEBUG oslo_concurrency.lockutils [req-66b1321f-dac2-41a0-9fcf-2f0867487e63 req-b830e666-9fc3-4e93-8495-aa2dd8dabc44 service nova] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.540290] env[65121]: DEBUG oslo_concurrency.lockutils [req-66b1321f-dac2-41a0-9fcf-2f0867487e63 req-b830e666-9fc3-4e93-8495-aa2dd8dabc44 service nova] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.540552] env[65121]: DEBUG oslo_concurrency.lockutils [req-66b1321f-dac2-41a0-9fcf-2f0867487e63 req-b830e666-9fc3-4e93-8495-aa2dd8dabc44 service nova] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.540791] env[65121]: DEBUG nova.compute.manager [req-66b1321f-dac2-41a0-9fcf-2f0867487e63 req-b830e666-9fc3-4e93-8495-aa2dd8dabc44 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] No waiting events found dispatching network-vif-plugged-2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1173.541040] env[65121]: WARNING nova.compute.manager [req-66b1321f-dac2-41a0-9fcf-2f0867487e63 req-b830e666-9fc3-4e93-8495-aa2dd8dabc44 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Received unexpected event network-vif-plugged-2dcc8fc2-ec38-4222-acf6-938b526306fa for instance with vm_state building and task_state spawning. [ 1173.663040] env[65121]: DEBUG nova.network.neutron [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Successfully updated port: 2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1173.676102] env[65121]: DEBUG oslo_vmware.api [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107399, 'name': ReconfigVM_Task, 'duration_secs': 5.784238} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.677132] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1173.677359] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Reconfigured VM to detach interface {{(pid=65121) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1173.677830] env[65121]: WARNING neutronclient.v2_0.client [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.678179] env[65121]: WARNING neutronclient.v2_0.client [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.678789] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.679230] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.689312] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1173.689651] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1173.689767] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1173.689938] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1173.690096] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1173.690241] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1173.690468] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1173.690593] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1173.690754] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1173.690914] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1173.691179] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1173.697410] env[65121]: WARNING neutronclient.v2_0.client [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.698161] env[65121]: WARNING openstack [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.698689] env[65121]: WARNING openstack [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.706125] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8df93fb-22ea-4daf-985f-6ef69c59e9f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.720835] env[65121]: WARNING neutronclient.v2_0.client [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.727136] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1173.727136] env[65121]: value = "task-5107412" [ 1173.727136] env[65121]: _type = "Task" [ 1173.727136] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.743500] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107412, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.809286] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107411, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.880852] env[65121]: DEBUG nova.compute.manager [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1173.880852] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1173.881789] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49deb5e8-9168-4c4a-9b3a-1954a8db3bcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.890465] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1173.891892] env[65121]: WARNING openstack [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.892280] env[65121]: WARNING openstack [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.898734] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a571000-3aca-4b1a-970f-22a4c3ff30ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.908164] env[65121]: DEBUG oslo_vmware.api [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1173.908164] env[65121]: value = "task-5107413" [ 1173.908164] env[65121]: _type = "Task" [ 1173.908164] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.921333] env[65121]: DEBUG oslo_vmware.api [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107413, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.970815] env[65121]: DEBUG nova.compute.manager [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1173.971134] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1173.971990] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c15314-89e8-4a0a-8815-7aafc341a623 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.981483] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1173.981786] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bec4f0c8-52ef-4236-9e7b-7cb94e069c12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.989179] env[65121]: DEBUG oslo_vmware.api [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1173.989179] env[65121]: value = "task-5107414" [ 1173.989179] env[65121]: _type = "Task" [ 1173.989179] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.999216] env[65121]: DEBUG oslo_vmware.api [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.037028] env[65121]: WARNING neutronclient.v2_0.client [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.037749] env[65121]: WARNING openstack [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.038114] env[65121]: WARNING openstack [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.055623] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.978s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.062638] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.466s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.062908] env[65121]: DEBUG nova.objects.instance [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lazy-loading 'resources' on Instance uuid f85ae5c9-ce3f-4ade-b708-4eae79485b20 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.171061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.171061] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1174.171061] env[65121]: DEBUG nova.network.neutron [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1174.187413] env[65121]: DEBUG nova.network.neutron [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Updating instance_info_cache with network_info: [{"id": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "address": "fa:16:3e:eb:2a:3d", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10c1c247-c8", "ovs_interfaceid": "10c1c247-c89e-4cd4-9ef3-0bf3621ad21c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1174.242042] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107412, 'name': ReconfigVM_Task, 'duration_secs': 0.359742} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.242390] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance '9a248b47-a34b-4a98-baf0-aa971b5aca94' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1174.312666] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107411, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546808} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.313146] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 138e48fe-0221-4173-801b-ee53b2bb98df/138e48fe-0221-4173-801b-ee53b2bb98df.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1174.313369] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1174.313649] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-67fe5b14-d516-4d69-be9e-a44ded4a89ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.322717] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1174.322717] env[65121]: value = "task-5107415" [ 1174.322717] env[65121]: _type = "Task" [ 1174.322717] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.334246] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.424084] env[65121]: DEBUG oslo_vmware.api [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107413, 'name': PowerOffVM_Task, 'duration_secs': 0.422097} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.424907] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.425664] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1174.426975] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79cfd3fb-7984-439a-be0a-d62ace0e0463 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.501028] env[65121]: DEBUG oslo_vmware.api [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107414, 'name': PowerOffVM_Task, 'duration_secs': 0.46751} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.501338] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.501498] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1174.501756] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da405fcb-2eaf-4da6-aa1f-fdd8c0df00b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.524186] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1174.524466] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1174.524670] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleting the datastore file [datastore2] 9475863f-86d5-44e6-ac19-93461ab87743 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.525020] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43042a81-382b-446a-9eca-a85aaee7b60e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.533170] env[65121]: DEBUG oslo_vmware.api [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for the task: (returnval){ [ 1174.533170] env[65121]: value = "task-5107418" [ 1174.533170] env[65121]: _type = "Task" [ 1174.533170] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.543358] env[65121]: DEBUG oslo_vmware.api [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107418, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.588401] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1174.588680] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1174.588914] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleting the datastore file [datastore2] 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.589279] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-759e4788-669e-4897-a7f4-ecda0081222a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.599622] env[65121]: DEBUG oslo_vmware.api [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1174.599622] env[65121]: value = "task-5107419" [ 1174.599622] env[65121]: _type = "Task" [ 1174.599622] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.613697] env[65121]: DEBUG oslo_vmware.api [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.675522] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.676019] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.691536] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1174.694050] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.694522] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.694522] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.694674] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.694798] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.699514] env[65121]: INFO nova.compute.manager [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Terminating instance [ 1174.721611] env[65121]: DEBUG nova.network.neutron [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1174.749732] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1174.749983] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1174.750154] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1174.750650] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1174.750650] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1174.750650] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1174.750816] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1174.750920] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1174.751098] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1174.751321] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1174.751493] env[65121]: DEBUG nova.virt.hardware [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1174.757140] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Reconfiguring VM instance instance-0000006f to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1174.761581] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.761973] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.769943] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1ea072a-7c64-462d-aa26-391d98ee1952 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.792204] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1174.792204] env[65121]: value = "task-5107420" [ 1174.792204] env[65121]: _type = "Task" [ 1174.792204] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.805950] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107420, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.840987] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080306} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.841544] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1174.842124] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cc6b02-050e-4b99-b898-f530d329a95a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.869531] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 138e48fe-0221-4173-801b-ee53b2bb98df/138e48fe-0221-4173-801b-ee53b2bb98df.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.871095] env[65121]: WARNING neutronclient.v2_0.client [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1174.871739] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1174.872090] env[65121]: WARNING openstack [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.882024] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb96e0e6-651a-4227-91e5-341a49ef9af1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.903724] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1174.903724] env[65121]: value = "task-5107421" [ 1174.903724] env[65121]: _type = "Task" [ 1174.903724] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.914684] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107421, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.940624] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a827f884-47d3-45fd-9478-7f592226d66a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.951178] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd01e5f-8423-405d-b932-6850483e8d60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.990094] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524174e8-4ce7-461c-839d-6cc6e97c4924 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.998754] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6600cbd4-f1d2-446b-b7a9-6b2cf34b4c42 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.005993] env[65121]: DEBUG nova.network.neutron [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updating instance_info_cache with network_info: [{"id": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "address": "fa:16:3e:28:94:f4", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dcc8fc2-ec", "ovs_interfaceid": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1175.016976] env[65121]: DEBUG nova.compute.provider_tree [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.044424] env[65121]: DEBUG oslo_vmware.api [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Task: {'id': task-5107418, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.235704} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.044681] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1175.044857] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1175.045043] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1175.045217] env[65121]: INFO nova.compute.manager [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1175.045511] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1175.045720] env[65121]: DEBUG nova.compute.manager [-] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1175.045821] env[65121]: DEBUG nova.network.neutron [-] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1175.046089] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.046813] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.047153] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.086604] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.103839] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.104099] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquired lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.104223] env[65121]: DEBUG nova.network.neutron [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1175.116313] env[65121]: DEBUG oslo_vmware.api [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217453} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.116976] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1175.117186] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1175.117529] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1175.117529] env[65121]: INFO nova.compute.manager [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1175.117743] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1175.117948] env[65121]: DEBUG nova.compute.manager [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1175.118223] env[65121]: DEBUG nova.network.neutron [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1175.118427] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.118945] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.119281] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.205134] env[65121]: DEBUG nova.compute.manager [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1175.205368] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1175.206390] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06679c04-8132-4079-95a4-cf0e3b334484 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.217509] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1175.217843] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65e22214-1cdf-48b7-9759-4687e5baf2d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.231082] env[65121]: DEBUG oslo_vmware.api [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1175.231082] env[65121]: value = "task-5107422" [ 1175.231082] env[65121]: _type = "Task" [ 1175.231082] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.241937] env[65121]: DEBUG oslo_vmware.api [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.250965] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.302813] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107420, 'name': ReconfigVM_Task, 'duration_secs': 0.189476} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.305011] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Reconfigured VM instance instance-0000006f to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1175.306177] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699be7d2-4cc8-4f96-8d09-7f8139a635c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.339312] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 9a248b47-a34b-4a98-baf0-aa971b5aca94/9a248b47-a34b-4a98-baf0-aa971b5aca94.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1175.340416] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-084d109b-515b-4813-b23e-f85a7b3663da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.363349] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1175.363349] env[65121]: value = "task-5107423" [ 1175.363349] env[65121]: _type = "Task" [ 1175.363349] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.376819] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107423, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.416641] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107421, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.441779] env[65121]: DEBUG nova.compute.manager [req-504b8129-fa35-4dec-90fc-e1a968b8194b req-81d8df3d-f332-40af-aba0-47fff5473b6f service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Received event network-vif-deleted-43502d20-e31f-4bc3-b5ff-4be5428a73a2 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1175.441779] env[65121]: INFO nova.compute.manager [req-504b8129-fa35-4dec-90fc-e1a968b8194b req-81d8df3d-f332-40af-aba0-47fff5473b6f service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Neutron deleted interface 43502d20-e31f-4bc3-b5ff-4be5428a73a2; detaching it from the instance and deleting it from the info cache [ 1175.442105] env[65121]: DEBUG nova.network.neutron [req-504b8129-fa35-4dec-90fc-e1a968b8194b req-81d8df3d-f332-40af-aba0-47fff5473b6f service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1175.522023] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1175.522023] env[65121]: DEBUG nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Instance network_info: |[{"id": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "address": "fa:16:3e:28:94:f4", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dcc8fc2-ec", "ovs_interfaceid": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1175.522023] env[65121]: DEBUG nova.scheduler.client.report [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1175.527697] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:94:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2dcc8fc2-ec38-4222-acf6-938b526306fa', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1175.538493] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1175.539789] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1175.540279] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e7cbc62-dfd6-4c7f-93d3-ff3a82aae588 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.564743] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1175.564743] env[65121]: value = "task-5107424" [ 1175.564743] env[65121]: _type = "Task" [ 1175.564743] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.574116] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107424, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.610999] env[65121]: INFO nova.compute.manager [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Swapping old allocation on dict_keys(['d56783bf-3ede-475a-8c5a-8d8303049e47']) held by migration c73f3217-4588-4e9d-b62a-5adf5dc65cf2 for instance [ 1175.612451] env[65121]: WARNING neutronclient.v2_0.client [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.613896] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.613896] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.650256] env[65121]: DEBUG nova.scheduler.client.report [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Overwriting current allocation {'allocations': {'d56783bf-3ede-475a-8c5a-8d8303049e47': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 161}}, 'project_id': '4467da3ed41245ddbc93fc865a8b7bdd', 'user_id': '63bf1d576d7e44c3a8f99a75546efe52', 'consumer_generation': 1} on consumer 84ae6061-a962-4a74-8661-9718cc4c5346 {{(pid=65121) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1175.659958] env[65121]: DEBUG nova.compute.manager [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Received event network-changed-2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1175.660252] env[65121]: DEBUG nova.compute.manager [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Refreshing instance network info cache due to event network-changed-2dcc8fc2-ec38-4222-acf6-938b526306fa. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1175.660539] env[65121]: DEBUG oslo_concurrency.lockutils [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Acquiring lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.660845] env[65121]: DEBUG oslo_concurrency.lockutils [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Acquired lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.661042] env[65121]: DEBUG nova.network.neutron [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Refreshing network info cache for port 2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1175.734165] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.749131] env[65121]: DEBUG oslo_vmware.api [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107422, 'name': PowerOffVM_Task, 'duration_secs': 0.315104} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.750280] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1175.750567] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1175.751226] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7df2e866-ff6b-4e9e-a761-bcbbbd837511 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.773904] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.774105] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.774278] env[65121]: DEBUG nova.network.neutron [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1175.809212] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.809544] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.817543] env[65121]: DEBUG nova.network.neutron [-] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1175.834665] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1175.834915] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1175.835085] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleting the datastore file [datastore2] 8fce40bb-69c1-4fbb-9b0b-c7aece54d179 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1175.835369] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e42060de-042b-40a4-a81f-06df5f83bef7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.847756] env[65121]: DEBUG oslo_vmware.api [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1175.847756] env[65121]: value = "task-5107426" [ 1175.847756] env[65121]: _type = "Task" [ 1175.847756] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.856642] env[65121]: DEBUG oslo_vmware.api [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107426, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.874433] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107423, 'name': ReconfigVM_Task, 'duration_secs': 0.335515} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.874750] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 9a248b47-a34b-4a98-baf0-aa971b5aca94/9a248b47-a34b-4a98-baf0-aa971b5aca94.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.877980] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance '9a248b47-a34b-4a98-baf0-aa971b5aca94' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1175.917992] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107421, 'name': ReconfigVM_Task, 'duration_secs': 0.79275} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.917992] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 138e48fe-0221-4173-801b-ee53b2bb98df/138e48fe-0221-4173-801b-ee53b2bb98df.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.919845] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d8d5ba9-edd3-4c82-b857-d5a5dc88e076 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.926602] env[65121]: WARNING neutronclient.v2_0.client [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.927347] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.927679] env[65121]: WARNING openstack [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.938707] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1175.938707] env[65121]: value = "task-5107427" [ 1175.938707] env[65121]: _type = "Task" [ 1175.938707] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.946046] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf985b6f-5277-4e3f-a099-108ea99f7a4a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.951871] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107427, 'name': Rename_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.962213] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f7d5e1-ceb9-49f5-bf7d-1d560aadd369 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.005255] env[65121]: DEBUG nova.compute.manager [req-504b8129-fa35-4dec-90fc-e1a968b8194b req-81d8df3d-f332-40af-aba0-47fff5473b6f service nova] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Detach interface failed, port_id=43502d20-e31f-4bc3-b5ff-4be5428a73a2, reason: Instance 9475863f-86d5-44e6-ac19-93461ab87743 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1176.040552] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.978s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.054948] env[65121]: INFO nova.network.neutron [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Port f6ae8a69-7565-42c4-96fe-1fbd5e038ebf from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1176.055362] env[65121]: DEBUG nova.network.neutron [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [{"id": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "address": "fa:16:3e:85:f6:68", "network": {"id": "5fdc1c3e-9b49-49d2-8a6f-f9c9467c27cd", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-59679449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35312a302644426f98f127e89a067e75", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1a1ddf-07", "ovs_interfaceid": "ff1a1ddf-0705-4111-80a1-faa79fe7c947", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1176.079661] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107424, 'name': CreateVM_Task, 'duration_secs': 0.422947} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.080881] env[65121]: INFO nova.scheduler.client.report [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted allocations for instance f85ae5c9-ce3f-4ade-b708-4eae79485b20 [ 1176.083189] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1176.086268] env[65121]: WARNING neutronclient.v2_0.client [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.086688] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.086866] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.087232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1176.088015] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-780b9837-5d14-4a34-9480-01de6bb71657 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.094759] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1176.094759] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a80b75-acf3-1188-a21a-c7202c236590" [ 1176.094759] env[65121]: _type = "Task" [ 1176.094759] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.107575] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a80b75-acf3-1188-a21a-c7202c236590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.165043] env[65121]: WARNING neutronclient.v2_0.client [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.165328] env[65121]: WARNING openstack [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.165628] env[65121]: WARNING openstack [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.252498] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1176.252845] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15be60e3-2ea9-4968-8c68-62d7dd56952c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.261788] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1176.261788] env[65121]: value = "task-5107428" [ 1176.261788] env[65121]: _type = "Task" [ 1176.261788] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.266488] env[65121]: WARNING openstack [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.266812] env[65121]: WARNING openstack [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.277292] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.278078] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.278438] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.289646] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107428, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.320690] env[65121]: INFO nova.compute.manager [-] [instance: 9475863f-86d5-44e6-ac19-93461ab87743] Took 1.27 seconds to deallocate network for instance. [ 1176.358925] env[65121]: DEBUG oslo_vmware.api [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107426, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.255284} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.360448] env[65121]: WARNING neutronclient.v2_0.client [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.361289] env[65121]: WARNING openstack [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.361711] env[65121]: WARNING openstack [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.371985] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1176.372245] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1176.372457] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1176.372625] env[65121]: INFO nova.compute.manager [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1176.372872] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1176.378498] env[65121]: DEBUG nova.compute.manager [-] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1176.378598] env[65121]: DEBUG nova.network.neutron [-] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1176.378866] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.379407] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.379660] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.389245] env[65121]: DEBUG nova.network.neutron [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1176.391914] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1985b0-0d68-4f5d-b1ce-5fff9f115d52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.420167] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2c0e09-09be-4ba9-9b98-9b673ca5dfb6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.448590] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance '9a248b47-a34b-4a98-baf0-aa971b5aca94' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1176.460099] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.474026] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107427, 'name': Rename_Task, 'duration_secs': 0.218464} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.474026] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1176.476039] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72e1a5da-1505-4510-98a4-a63debf5be51 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.484214] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1176.484214] env[65121]: value = "task-5107429" [ 1176.484214] env[65121]: _type = "Task" [ 1176.484214] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.494059] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.495929] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.496073] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.552499] env[65121]: DEBUG nova.network.neutron [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updated VIF entry in instance network info cache for port 2dcc8fc2-ec38-4222-acf6-938b526306fa. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1176.553751] env[65121]: DEBUG nova.network.neutron [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updating instance_info_cache with network_info: [{"id": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "address": "fa:16:3e:28:94:f4", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dcc8fc2-ec", "ovs_interfaceid": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1176.558864] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Releasing lock "refresh_cache-8fce40bb-69c1-4fbb-9b0b-c7aece54d179" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1176.591997] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9b66ea60-a62b-49a5-a1e8-db8412522cdc tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f85ae5c9-ce3f-4ade-b708-4eae79485b20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.491s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.611036] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a80b75-acf3-1188-a21a-c7202c236590, 'name': SearchDatastore_Task, 'duration_secs': 0.015211} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.615718] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1176.615989] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1176.616277] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.616491] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.616644] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.617603] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c257b292-6e7b-4029-bfb5-ee31556a6a25 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.629533] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.629855] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1176.630565] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-746d4faf-c2db-47ef-bbd4-c8f3e605754a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.638613] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1176.638613] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]521474cf-5583-e64e-a828-075a9169cee8" [ 1176.638613] env[65121]: _type = "Task" [ 1176.638613] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.640180] env[65121]: WARNING neutronclient.v2_0.client [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.640903] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.641314] env[65121]: WARNING openstack [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.660252] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521474cf-5583-e64e-a828-075a9169cee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.775870] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107428, 'name': PowerOffVM_Task, 'duration_secs': 0.219507} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.775870] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1176.776821] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc8b980-dc97-4ee3-89bb-c866dcf248d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.800665] env[65121]: DEBUG nova.network.neutron [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [{"id": "2acd5dbc-a08c-4ebd-922b-284294369a33", "address": "fa:16:3e:75:84:87", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2acd5dbc-a0", "ovs_interfaceid": "2acd5dbc-a08c-4ebd-922b-284294369a33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1176.802555] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ca2fa9-170c-43d0-9bc3-3dc4b974ea4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.828572] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.829351] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.829783] env[65121]: DEBUG nova.objects.instance [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lazy-loading 'resources' on Instance uuid 9475863f-86d5-44e6-ac19-93461ab87743 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.846573] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1176.846895] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c57f211-a25c-4b47-8fae-632e03b5822b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.855141] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1176.855141] env[65121]: value = "task-5107430" [ 1176.855141] env[65121]: _type = "Task" [ 1176.855141] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.865518] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1176.865769] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1176.866083] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.866252] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.866437] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.866746] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9125440-b0ec-4da7-894f-c0819c62ec75 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.879376] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.879653] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1176.880450] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-898ec46c-a79e-427b-812a-56041ac05575 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.887999] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1176.887999] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]523b4203-ce38-7730-8e19-2b1bccc32d1b" [ 1176.887999] env[65121]: _type = "Task" [ 1176.887999] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.897661] env[65121]: INFO nova.compute.manager [-] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Took 1.78 seconds to deallocate network for instance. [ 1176.897997] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523b4203-ce38-7730-8e19-2b1bccc32d1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.969413] env[65121]: WARNING neutronclient.v2_0.client [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.995188] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107429, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.014150] env[65121]: DEBUG nova.network.neutron [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Port d86b7885-6e5e-45ec-b64e-19ae24268b82 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1177.058250] env[65121]: DEBUG oslo_concurrency.lockutils [req-04d50789-6a0b-4485-b74f-2287382857b1 req-0ae9fc56-bff1-473d-8fc9-e61a90e9f894 service nova] Releasing lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.064108] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77ebbac4-91cd-4e19-a5fd-cd74ba235ee5 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "interface-8fce40bb-69c1-4fbb-9b0b-c7aece54d179-f6ae8a69-7565-42c4-96fe-1fbd5e038ebf" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.030s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.163585] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]521474cf-5583-e64e-a828-075a9169cee8, 'name': SearchDatastore_Task, 'duration_secs': 0.022469} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.164444] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb7c6075-48f7-46ed-ad57-7fcd35db343c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.172865] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1177.172865] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52349570-ba02-29c8-9a30-fe85032a2067" [ 1177.172865] env[65121]: _type = "Task" [ 1177.172865] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.187432] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52349570-ba02-29c8-9a30-fe85032a2067, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.307856] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-84ae6061-a962-4a74-8661-9718cc4c5346" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.308907] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399925c9-8ec6-4f9e-97dd-900837c3d662 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.320509] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bdaff5-eef4-4eef-8021-abbbce953ba0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.402741] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]523b4203-ce38-7730-8e19-2b1bccc32d1b, 'name': SearchDatastore_Task, 'duration_secs': 0.016036} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.403224] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a497a47-c582-43e9-b18f-ce5021aaf53f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.407247] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.415130] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1177.415130] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5247e411-ed51-435b-1722-b90d7819e598" [ 1177.415130] env[65121]: _type = "Task" [ 1177.415130] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.427203] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5247e411-ed51-435b-1722-b90d7819e598, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.499302] env[65121]: DEBUG oslo_vmware.api [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107429, 'name': PowerOnVM_Task, 'duration_secs': 0.593685} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.499565] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1177.499771] env[65121]: INFO nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Took 9.35 seconds to spawn the instance on the hypervisor. [ 1177.499990] env[65121]: DEBUG nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1177.501039] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cae4d2-268c-4a07-b761-86778067c86e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.623092] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ae3e98-f08b-475e-ad41-1b7e60c37529 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.632226] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcd2843-ab9e-4046-bc72-e7d410add4df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.667372] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a715f5f1-b725-4150-ad74-30b5efc3d221 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.677889] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90925c55-29eb-4477-bc1e-e745e54f9f56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.692332] env[65121]: DEBUG nova.compute.provider_tree [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.697213] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52349570-ba02-29c8-9a30-fe85032a2067, 'name': SearchDatastore_Task, 'duration_secs': 0.018318} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.697997] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.698270] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2/b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1177.698534] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad8b6ed8-2866-4f5d-8151-e753de55b098 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.706873] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1177.706873] env[65121]: value = "task-5107431" [ 1177.706873] env[65121]: _type = "Task" [ 1177.706873] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.717859] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.762776] env[65121]: DEBUG nova.compute.manager [req-6fd5a68c-eb4e-4687-a204-5c7428a95987 req-f6c73ba6-1f6b-479b-8a27-fe240cd26caa service nova] [instance: 1fbd3a12-82d3-458d-a590-0f5c0c70ae55] Received event network-vif-deleted-53616727-ab6a-4f06-a597-0fc068a18955 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1177.762989] env[65121]: DEBUG nova.compute.manager [req-6fd5a68c-eb4e-4687-a204-5c7428a95987 req-f6c73ba6-1f6b-479b-8a27-fe240cd26caa service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Received event network-vif-deleted-ff1a1ddf-0705-4111-80a1-faa79fe7c947 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1177.763465] env[65121]: INFO nova.compute.manager [req-6fd5a68c-eb4e-4687-a204-5c7428a95987 req-f6c73ba6-1f6b-479b-8a27-fe240cd26caa service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Neutron deleted interface ff1a1ddf-0705-4111-80a1-faa79fe7c947; detaching it from the instance and deleting it from the info cache [ 1177.763640] env[65121]: DEBUG nova.network.neutron [req-6fd5a68c-eb4e-4687-a204-5c7428a95987 req-f6c73ba6-1f6b-479b-8a27-fe240cd26caa service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1177.844636] env[65121]: DEBUG nova.network.neutron [None req-d9e5242a-3944-45c4-a6b9-faec961297a5 None None] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1177.929681] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5247e411-ed51-435b-1722-b90d7819e598, 'name': SearchDatastore_Task, 'duration_secs': 0.014298} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.929923] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.930220] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. {{(pid=65121) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1177.930511] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bff48c8f-4ddb-4f35-b000-f57d74235cf6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.940637] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1177.940637] env[65121]: value = "task-5107432" [ 1177.940637] env[65121]: _type = "Task" [ 1177.940637] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.950706] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.044429] env[65121]: INFO nova.compute.manager [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Took 21.53 seconds to build instance. [ 1178.051271] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.051654] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.051902] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.101331] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.101575] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.101775] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.101948] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.102125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.104536] env[65121]: INFO nova.compute.manager [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Terminating instance [ 1178.199362] env[65121]: DEBUG nova.scheduler.client.report [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1178.218137] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49644} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.218267] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2/b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1178.218478] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1178.219478] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f04ed21b-24f4-4685-9650-4f0d27f40619 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.227633] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1178.227633] env[65121]: value = "task-5107433" [ 1178.227633] env[65121]: _type = "Task" [ 1178.227633] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.238630] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107433, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.267045] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-048c623b-0212-4c1f-a5cf-ffdaad7c62fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.277503] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5d0d69-a8e8-458e-82f5-7a4b9b81ced4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.317517] env[65121]: DEBUG nova.compute.manager [req-6fd5a68c-eb4e-4687-a204-5c7428a95987 req-f6c73ba6-1f6b-479b-8a27-fe240cd26caa service nova] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Detach interface failed, port_id=ff1a1ddf-0705-4111-80a1-faa79fe7c947, reason: Instance 8fce40bb-69c1-4fbb-9b0b-c7aece54d179 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1178.347683] env[65121]: INFO nova.compute.manager [None req-d9e5242a-3944-45c4-a6b9-faec961297a5 None None] [instance: 8fce40bb-69c1-4fbb-9b0b-c7aece54d179] Took 1.97 seconds to deallocate network for instance. [ 1178.424101] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.424877] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5353f874-b1ae-4761-9abd-c8b7425b974f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.437183] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1178.437183] env[65121]: value = "task-5107434" [ 1178.437183] env[65121]: _type = "Task" [ 1178.437183] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.457233] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107432, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.457969] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.554932] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3a5b5d1f-0517-4a55-b845-34ba28121773 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.073s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.609037] env[65121]: DEBUG nova.compute.manager [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1178.609037] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1178.610067] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0365fd37-a165-41e7-91f0-661d29c07db0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.619073] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.619346] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-563f182e-f1ee-43b6-a1e0-0db037d27ce5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.627042] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1178.627042] env[65121]: value = "task-5107435" [ 1178.627042] env[65121]: _type = "Task" [ 1178.627042] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.635653] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.705948] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.708462] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.301s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.708750] env[65121]: DEBUG nova.objects.instance [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'resources' on Instance uuid 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1178.735755] env[65121]: INFO nova.scheduler.client.report [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Deleted allocations for instance 9475863f-86d5-44e6-ac19-93461ab87743 [ 1178.743326] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107433, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06964} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.743590] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1178.744405] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96dff66-6faa-4402-9416-b8347e84be43 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.772597] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2/b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1178.772920] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a1889ff-9284-474a-b251-51de81c05250 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.799083] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1178.799083] env[65121]: value = "task-5107436" [ 1178.799083] env[65121]: _type = "Task" [ 1178.799083] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.810507] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107436, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.855404] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.953025] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107434, 'name': PowerOffVM_Task, 'duration_secs': 0.483034} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.958719] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.958719] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1178.958719] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1178.959171] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1178.959439] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1178.959639] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1178.959870] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1178.960242] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1178.960474] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1178.960694] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1178.960913] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1178.961169] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1178.966707] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107432, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.966990] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd67b8f2-955e-48ee-8e77-5bc1de2c18ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.986158] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1178.986158] env[65121]: value = "task-5107437" [ 1178.986158] env[65121]: _type = "Task" [ 1178.986158] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.001021] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107437, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.059785] env[65121]: WARNING neutronclient.v2_0.client [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.139510] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.141298] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.141494] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1179.141641] env[65121]: DEBUG nova.network.neutron [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1179.217074] env[65121]: DEBUG nova.compute.manager [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Received event network-changed-3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1179.217789] env[65121]: DEBUG nova.compute.manager [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Refreshing instance network info cache due to event network-changed-3987861c-5f58-4f3f-868e-31452ab11d4d. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1179.217944] env[65121]: DEBUG oslo_concurrency.lockutils [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Acquiring lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.218104] env[65121]: DEBUG oslo_concurrency.lockutils [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Acquired lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1179.218505] env[65121]: DEBUG nova.network.neutron [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Refreshing network info cache for port 3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1179.247576] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ca48f0fb-a1ad-4b36-8fa5-35e16df360cc tempest-ServerDiskConfigTestJSON-1517940131 tempest-ServerDiskConfigTestJSON-1517940131-project-member] Lock "9475863f-86d5-44e6-ac19-93461ab87743" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.878s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.313407] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.459933] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107432, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.469455] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c39cba-ecec-4e85-8b9c-032f5bed7df0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.481114] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c149f7af-fc29-499b-8182-4ae73c1ccc3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.499459] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107437, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.528505] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fa973e-0ec2-4cce-ae71-fcbfff9454cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.537138] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072e0178-befb-439b-ac73-251956fe3ee9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.552759] env[65121]: DEBUG nova.compute.provider_tree [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1179.640693] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.644655] env[65121]: WARNING neutronclient.v2_0.client [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.645531] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.645934] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.722234] env[65121]: WARNING neutronclient.v2_0.client [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.723396] env[65121]: WARNING openstack [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.723942] env[65121]: WARNING openstack [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.796697] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.797539] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.819715] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.918328] env[65121]: WARNING openstack [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.918768] env[65121]: WARNING openstack [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.944870] env[65121]: WARNING neutronclient.v2_0.client [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.945657] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.946023] env[65121]: WARNING openstack [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.964086] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107432, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.666909} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.964358] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. [ 1179.966452] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e948af-a8cd-4411-97e5-3fe578928748 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.009215] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.017050] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8708f3bc-6785-466a-aac9-30d6d1776624 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.033703] env[65121]: WARNING neutronclient.v2_0.client [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1180.034593] env[65121]: WARNING openstack [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1180.035077] env[65121]: WARNING openstack [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1180.047943] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107437, 'name': ReconfigVM_Task, 'duration_secs': 0.549092} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.048221] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1180.048221] env[65121]: value = "task-5107438" [ 1180.048221] env[65121]: _type = "Task" [ 1180.048221] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.049246] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91337db7-7227-4394-a337-2b4e875fdeba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.058135] env[65121]: DEBUG nova.scheduler.client.report [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1180.068031] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.087872] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1180.087872] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1180.087872] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1180.087872] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1180.087872] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1180.087872] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1180.088391] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1180.088391] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1180.088391] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1180.088527] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1180.088739] env[65121]: DEBUG nova.virt.hardware [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1180.089967] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33af5d1-2ab4-4394-9b64-7d16a03caefb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.096996] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1180.096996] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c69480-5b3b-24dd-0b90-3de8d74d767b" [ 1180.096996] env[65121]: _type = "Task" [ 1180.096996] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.109025] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c69480-5b3b-24dd-0b90-3de8d74d767b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.139690] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107435, 'name': PowerOffVM_Task, 'duration_secs': 1.23443} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.140109] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1180.140278] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1180.140539] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3cf34ec-8919-4192-af07-20fd72018862 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.163413] env[65121]: DEBUG nova.network.neutron [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance_info_cache with network_info: [{"id": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "address": "fa:16:3e:22:e8:da", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd86b7885-6e", "ovs_interfaceid": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1180.217562] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1180.217824] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1180.218013] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleting the datastore file [datastore2] f84d7aa1-a90b-4c95-9673-fb1a2af7187c {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1180.218337] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f8d8b3e-ae8a-43b4-a2f9-9ad165d78b96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.227892] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for the task: (returnval){ [ 1180.227892] env[65121]: value = "task-5107440" [ 1180.227892] env[65121]: _type = "Task" [ 1180.227892] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.238914] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107440, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.303088] env[65121]: DEBUG nova.network.neutron [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Updated VIF entry in instance network info cache for port 3987861c-5f58-4f3f-868e-31452ab11d4d. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1180.303259] env[65121]: DEBUG nova.network.neutron [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Updating instance_info_cache with network_info: [{"id": "3987861c-5f58-4f3f-868e-31452ab11d4d", "address": "fa:16:3e:75:a0:1a", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3987861c-5f", "ovs_interfaceid": "3987861c-5f58-4f3f-868e-31452ab11d4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1180.319371] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107436, 'name': ReconfigVM_Task, 'duration_secs': 1.200749} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.320383] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfigured VM instance instance-00000075 to attach disk [datastore1] b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2/b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.321108] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f21306c-9fbc-4b53-82ea-08380e7e05d1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.331202] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1180.331202] env[65121]: value = "task-5107441" [ 1180.331202] env[65121]: _type = "Task" [ 1180.331202] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.341625] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107441, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.562958] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107438, 'name': ReconfigVM_Task, 'duration_secs': 0.473602} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.562958] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Reconfigured VM instance instance-00000073 to attach disk [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.564612] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9263bb-c203-4a81-8805-9ae997134126 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.567876] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.859s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.570662] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.715s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1180.570905] env[65121]: DEBUG nova.objects.instance [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'resources' on Instance uuid 8fce40bb-69c1-4fbb-9b0b-c7aece54d179 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1180.612803] env[65121]: INFO nova.scheduler.client.report [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted allocations for instance 1fbd3a12-82d3-458d-a590-0f5c0c70ae55 [ 1180.620771] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dc56973-9322-4975-ac7f-94db320bbe31 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.647025] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c69480-5b3b-24dd-0b90-3de8d74d767b, 'name': SearchDatastore_Task, 'duration_secs': 0.009574} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.653197] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1180.653602] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1180.653602] env[65121]: value = "task-5107442" [ 1180.653602] env[65121]: _type = "Task" [ 1180.653602] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.654215] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e92e905-85e7-4174-95d0-d9c18009a282 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.673086] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1180.689945] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107442, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.691690] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1180.691690] env[65121]: value = "task-5107443" [ 1180.691690] env[65121]: _type = "Task" [ 1180.691690] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.705019] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107443, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.739720] env[65121]: DEBUG oslo_vmware.api [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Task: {'id': task-5107440, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205832} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.739924] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1180.740034] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1180.741263] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1180.741263] env[65121]: INFO nova.compute.manager [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Took 2.13 seconds to destroy the instance on the hypervisor. [ 1180.741263] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1180.741263] env[65121]: DEBUG nova.compute.manager [-] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1180.741263] env[65121]: DEBUG nova.network.neutron [-] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1180.741796] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1180.742106] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1180.742315] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1180.785360] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1180.808555] env[65121]: DEBUG oslo_concurrency.lockutils [req-959fe0a5-8247-4847-afef-3d481111273d req-0b479cd9-eb3d-4404-a8fb-ad1650870349 service nova] Releasing lock "refresh_cache-138e48fe-0221-4173-801b-ee53b2bb98df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1180.844769] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107441, 'name': Rename_Task, 'duration_secs': 0.16793} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.845155] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1180.845431] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d94c4c07-de06-4329-8ab9-f8aeda4eaf3d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.857445] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1180.857445] env[65121]: value = "task-5107444" [ 1180.857445] env[65121]: _type = "Task" [ 1180.857445] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.870849] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.899193] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56d6533-8f09-4bc0-8ae3-f43ea9a3fae0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.907641] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabfb3a1-e83e-4b24-a261-92e063ccfee4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.945292] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf205c93-3519-437f-afe0-4fb65f7b64fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.954551] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc26df5-b744-47ce-8b23-f89b7d5ff788 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.971719] env[65121]: DEBUG nova.compute.provider_tree [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1181.139641] env[65121]: DEBUG oslo_concurrency.lockutils [None req-066dd0a8-eaa4-4d5c-b9b5-0141d86f4c45 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "1fbd3a12-82d3-458d-a590-0f5c0c70ae55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.679s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.180117] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107442, 'name': ReconfigVM_Task, 'duration_secs': 0.313608} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.180433] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1181.180633] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7aa3edaa-fcdf-42d8-946d-08f9c2d9e43d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.191914] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1181.191914] env[65121]: value = "task-5107445" [ 1181.191914] env[65121]: _type = "Task" [ 1181.191914] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.202791] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0093fc7-86ba-4aba-b5d8-17909fe71f6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.211660] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107443, 'name': ReconfigVM_Task, 'duration_secs': 0.342872} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.230447] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1181.230560] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107445, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.231391] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6052337-4737-4027-9c9e-af9ca30b5c50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.235074] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320da2c9-8069-423b-bdcd-a94a4aeb4999 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.245087] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance '9a248b47-a34b-4a98-baf0-aa971b5aca94' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1181.273558] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.275086] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f72d1483-d637-45e0-80d6-6fa10cbb2291 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.290523] env[65121]: DEBUG nova.compute.manager [req-d0900cdb-6588-4894-8f32-bdf61748db09 req-44da1c99-569b-4328-82b1-c164ec8f9c2f service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Received event network-vif-deleted-bb4123fb-a2e7-46f0-b5b1-73cd269f125d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1181.290523] env[65121]: INFO nova.compute.manager [req-d0900cdb-6588-4894-8f32-bdf61748db09 req-44da1c99-569b-4328-82b1-c164ec8f9c2f service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Neutron deleted interface bb4123fb-a2e7-46f0-b5b1-73cd269f125d; detaching it from the instance and deleting it from the info cache [ 1181.290730] env[65121]: DEBUG nova.network.neutron [req-d0900cdb-6588-4894-8f32-bdf61748db09 req-44da1c99-569b-4328-82b1-c164ec8f9c2f service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1181.301632] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1181.301632] env[65121]: value = "task-5107446" [ 1181.301632] env[65121]: _type = "Task" [ 1181.301632] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.313107] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107446, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.369737] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107444, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.475692] env[65121]: DEBUG nova.scheduler.client.report [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1181.524771] env[65121]: DEBUG nova.network.neutron [-] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1181.702696] env[65121]: DEBUG oslo_vmware.api [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107445, 'name': PowerOnVM_Task, 'duration_secs': 0.502265} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.708163] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1181.712818] env[65121]: DEBUG nova.compute.manager [None req-6caada4d-d794-4fe3-8a00-8bd9337e2598 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1181.713703] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462d0b0b-7846-400e-9687-35f4602ca966 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.779331] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1181.780974] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a159bc80-c9f5-4e2b-a524-193e1fb8bb65 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.791488] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1181.791488] env[65121]: value = "task-5107447" [ 1181.791488] env[65121]: _type = "Task" [ 1181.791488] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.796731] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd846f86-952f-4ad0-9e8d-63388ed0c4be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.811325] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.815313] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8152aea3-16d7-4f12-b17e-9feec72c73fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.833798] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107446, 'name': ReconfigVM_Task, 'duration_secs': 0.441732} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.834577] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346/84ae6061-a962-4a74-8661-9718cc4c5346.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1181.835560] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704a3c8d-5313-4919-b5d1-0d57e4b4b859 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.873215] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37c4a95-b3a0-4fc7-b215-818c66575a85 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.876741] env[65121]: DEBUG nova.compute.manager [req-d0900cdb-6588-4894-8f32-bdf61748db09 req-44da1c99-569b-4328-82b1-c164ec8f9c2f service nova] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Detach interface failed, port_id=bb4123fb-a2e7-46f0-b5b1-73cd269f125d, reason: Instance f84d7aa1-a90b-4c95-9673-fb1a2af7187c could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1181.893148] env[65121]: DEBUG oslo_vmware.api [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107444, 'name': PowerOnVM_Task, 'duration_secs': 0.7189} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.910837] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1181.911207] env[65121]: INFO nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Took 8.81 seconds to spawn the instance on the hypervisor. [ 1181.911390] env[65121]: DEBUG nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1181.912368] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d0fad5-7a9e-4cd4-924b-cdc5a97d3431 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.915639] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e12459-d313-41d8-b712-ec3eed4ef0ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.945759] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de16b5de-1e05-4640-9bf7-3f8c81f7293f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.954092] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1181.954092] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c455345e-1f2a-4cd1-9472-2080beb3d989 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.960441] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1181.960441] env[65121]: value = "task-5107448" [ 1181.960441] env[65121]: _type = "Task" [ 1181.960441] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.969543] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107448, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.982486] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.412s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.008237] env[65121]: INFO nova.scheduler.client.report [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted allocations for instance 8fce40bb-69c1-4fbb-9b0b-c7aece54d179 [ 1182.029416] env[65121]: INFO nova.compute.manager [-] [instance: f84d7aa1-a90b-4c95-9673-fb1a2af7187c] Took 1.29 seconds to deallocate network for instance. [ 1182.303658] env[65121]: DEBUG oslo_vmware.api [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107447, 'name': PowerOnVM_Task, 'duration_secs': 0.511946} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.303658] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1182.303804] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-4e3be0c5-b013-49b8-8475-86cae1dbe571 tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance '9a248b47-a34b-4a98-baf0-aa971b5aca94' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1182.456770] env[65121]: INFO nova.compute.manager [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Took 24.42 seconds to build instance. [ 1182.471802] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107448, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.519334] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9fa5e905-6d0a-4ba2-822e-d4bf91badc11 tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "8fce40bb-69c1-4fbb-9b0b-c7aece54d179" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.825s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.552479] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.552788] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.553020] env[65121]: DEBUG nova.objects.instance [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lazy-loading 'resources' on Instance uuid f84d7aa1-a90b-4c95-9673-fb1a2af7187c {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1182.959349] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d97b141b-0c5f-4d85-9b8a-76564e1a6038 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.959s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.977531] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107448, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.162599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "992ba328-206a-4d47-a28b-f3ef9b20817f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.163339] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.200810] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.201206] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.201338] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.201465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.201628] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.205123] env[65121]: INFO nova.compute.manager [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Terminating instance [ 1183.335776] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db12241d-b9d3-42c1-85ed-eba7d36a3005 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.347377] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce28386-4c37-41d0-b742-8a509d5dc93f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.387251] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf081951-844c-46df-97bb-708bd8f5d2a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.395144] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aaf974b-e0b1-43a9-950c-76f22400eca8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.414388] env[65121]: DEBUG nova.compute.provider_tree [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1183.478278] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107448, 'name': PowerOnVM_Task} progress is 74%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.517276] env[65121]: DEBUG nova.compute.manager [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Received event network-changed-2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1183.517276] env[65121]: DEBUG nova.compute.manager [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Refreshing instance network info cache due to event network-changed-2dcc8fc2-ec38-4222-acf6-938b526306fa. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1183.517276] env[65121]: DEBUG oslo_concurrency.lockutils [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Acquiring lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.517276] env[65121]: DEBUG oslo_concurrency.lockutils [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Acquired lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.517276] env[65121]: DEBUG nova.network.neutron [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Refreshing network info cache for port 2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1183.669149] env[65121]: DEBUG nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1183.711138] env[65121]: DEBUG nova.compute.manager [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1183.711449] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1183.712456] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7473a8f-bcc9-4309-941a-ac88f2a37356 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.722287] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1183.722557] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-557102bd-50d5-440d-afc4-7be137eded8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.729997] env[65121]: DEBUG oslo_vmware.api [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1183.729997] env[65121]: value = "task-5107449" [ 1183.729997] env[65121]: _type = "Task" [ 1183.729997] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.739351] env[65121]: DEBUG oslo_vmware.api [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.918516] env[65121]: DEBUG nova.scheduler.client.report [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1183.976689] env[65121]: DEBUG oslo_vmware.api [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107448, 'name': PowerOnVM_Task, 'duration_secs': 1.917297} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.977271] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1184.018194] env[65121]: WARNING neutronclient.v2_0.client [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1184.019335] env[65121]: WARNING openstack [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.019897] env[65121]: WARNING openstack [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.151729] env[65121]: WARNING openstack [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.152241] env[65121]: WARNING openstack [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.205695] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.224387] env[65121]: WARNING neutronclient.v2_0.client [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1184.225181] env[65121]: WARNING openstack [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.227124] env[65121]: WARNING openstack [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.247555] env[65121]: DEBUG oslo_vmware.api [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107449, 'name': PowerOffVM_Task, 'duration_secs': 0.378497} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.248779] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1184.248779] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1184.248779] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4395c787-3085-4b73-82c1-442f2a3ede49 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.334205] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1184.334205] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1184.334205] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleting the datastore file [datastore1] 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1184.337890] env[65121]: DEBUG nova.network.neutron [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updated VIF entry in instance network info cache for port 2dcc8fc2-ec38-4222-acf6-938b526306fa. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1184.337890] env[65121]: DEBUG nova.network.neutron [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updating instance_info_cache with network_info: [{"id": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "address": "fa:16:3e:28:94:f4", "network": {"id": "b588c8f3-4990-4a29-bcfd-da6ea3577715", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1145888305-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ce7f6698e214d73ae43427601058af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dcc8fc2-ec", "ovs_interfaceid": "2dcc8fc2-ec38-4222-acf6-938b526306fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1184.339644] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28ea43d4-1ab9-4d43-b84a-ecd667b1e25e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.349777] env[65121]: DEBUG oslo_vmware.api [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for the task: (returnval){ [ 1184.349777] env[65121]: value = "task-5107451" [ 1184.349777] env[65121]: _type = "Task" [ 1184.349777] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.359131] env[65121]: DEBUG oslo_vmware.api [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107451, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.427981] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.436561] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.227s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.438424] env[65121]: INFO nova.compute.claims [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1184.472789] env[65121]: INFO nova.scheduler.client.report [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Deleted allocations for instance f84d7aa1-a90b-4c95-9673-fb1a2af7187c [ 1184.843336] env[65121]: DEBUG oslo_concurrency.lockutils [req-91520f57-b37d-46bb-82d9-66a1c78af6f6 req-c8c56181-bb01-4cd2-8951-bc0db2fc20d2 service nova] Releasing lock "refresh_cache-b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.868637] env[65121]: DEBUG oslo_vmware.api [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Task: {'id': task-5107451, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163486} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.868637] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.869023] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1184.869023] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1184.870310] env[65121]: INFO nova.compute.manager [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1184.870989] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1184.871241] env[65121]: DEBUG nova.compute.manager [-] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1184.871365] env[65121]: DEBUG nova.network.neutron [-] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1184.871669] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1184.873763] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1184.873763] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1184.985360] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1184.993194] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5f854560-d51b-48b6-a070-b13dd8efe865 tempest-ServersTestJSON-1776500097 tempest-ServersTestJSON-1776500097-project-member] Lock "f84d7aa1-a90b-4c95-9673-fb1a2af7187c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.890s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1185.047840] env[65121]: INFO nova.compute.manager [None req-7e2329b7-8a85-4c7a-9bac-3d1130559bca tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance to original state: 'active' [ 1185.339740] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1185.340089] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1185.340710] env[65121]: DEBUG nova.compute.manager [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Going to confirm migration 8 {{(pid=65121) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1185.546629] env[65121]: INFO nova.compute.manager [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Rescuing [ 1185.546934] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.547230] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.547320] env[65121]: DEBUG nova.network.neutron [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1185.702815] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31048b57-3422-49eb-a919-61ac42077d0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.712389] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46add7a8-fd41-404e-8321-e38cb95306b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.756979] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597bf79a-08c3-48c0-9bd9-5135407b5bf0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.766219] env[65121]: DEBUG nova.network.neutron [-] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1185.769235] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23cab5f-5307-44b2-a8f0-02bb63503126 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.785889] env[65121]: DEBUG nova.compute.provider_tree [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1185.848959] env[65121]: WARNING neutronclient.v2_0.client [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1185.859731] env[65121]: DEBUG nova.compute.manager [req-68069569-118a-4861-8106-b23058731448 req-9557636a-75f3-41cf-ac5a-b83b9c178ed2 service nova] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Received event network-vif-deleted-9dc8b4e9-1442-43ba-a663-567f89e3a428 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1185.923537] env[65121]: WARNING neutronclient.v2_0.client [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1185.923715] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.923862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquired lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.924045] env[65121]: DEBUG nova.network.neutron [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1185.924230] env[65121]: DEBUG nova.objects.instance [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'info_cache' on Instance uuid 9a248b47-a34b-4a98-baf0-aa971b5aca94 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.052082] env[65121]: WARNING neutronclient.v2_0.client [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.052082] env[65121]: WARNING openstack [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.052082] env[65121]: WARNING openstack [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.225644] env[65121]: WARNING openstack [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.227176] env[65121]: WARNING openstack [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.275243] env[65121]: INFO nova.compute.manager [-] [instance: 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7] Took 1.40 seconds to deallocate network for instance. [ 1186.295917] env[65121]: DEBUG nova.scheduler.client.report [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1186.304822] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.307357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.349238] env[65121]: WARNING neutronclient.v2_0.client [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.350724] env[65121]: WARNING openstack [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.351205] env[65121]: WARNING openstack [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.507396] env[65121]: DEBUG nova.network.neutron [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Updating instance_info_cache with network_info: [{"id": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "address": "fa:16:3e:19:1a:cf", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf714bfb3-87", "ovs_interfaceid": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1186.782714] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.805846] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.806418] env[65121]: DEBUG nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1186.811829] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.027s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.811829] env[65121]: DEBUG nova.objects.instance [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lazy-loading 'resources' on Instance uuid 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.811829] env[65121]: DEBUG nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1186.931654] env[65121]: WARNING neutronclient.v2_0.client [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.933293] env[65121]: WARNING openstack [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.933645] env[65121]: WARNING openstack [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.010422] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.062618] env[65121]: WARNING openstack [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.063018] env[65121]: WARNING openstack [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.072858] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "84ae6061-a962-4a74-8661-9718cc4c5346" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.072995] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.073230] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.073413] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.073944] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1187.076060] env[65121]: INFO nova.compute.manager [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Terminating instance [ 1187.129367] env[65121]: WARNING neutronclient.v2_0.client [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.130058] env[65121]: WARNING openstack [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.130402] env[65121]: WARNING openstack [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.219729] env[65121]: DEBUG nova.network.neutron [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance_info_cache with network_info: [{"id": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "address": "fa:16:3e:22:e8:da", "network": {"id": "a4ff8134-dcbb-4e42-9d1c-f09ddfa7a154", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-431352460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0630960dcbf44781be05184565d81932", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd86b7885-6e", "ovs_interfaceid": "d86b7885-6e5e-45ec-b64e-19ae24268b82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1187.315696] env[65121]: DEBUG nova.compute.utils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1187.322340] env[65121]: DEBUG nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1187.322571] env[65121]: DEBUG nova.network.neutron [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1187.323137] env[65121]: WARNING neutronclient.v2_0.client [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.323664] env[65121]: WARNING neutronclient.v2_0.client [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.324599] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.325261] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.355644] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.384585] env[65121]: DEBUG nova.policy [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8931a266fddc4346b0c5243608a39bff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00383aa3355e438cb703c2b86c7917f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1187.580361] env[65121]: DEBUG nova.compute.manager [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1187.580604] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.580891] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b572de55-597f-4746-b52e-d0840c887b8f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.586247] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40da19be-0ac6-4ed4-9d06-89133e699b23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.590576] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1187.590576] env[65121]: value = "task-5107452" [ 1187.590576] env[65121]: _type = "Task" [ 1187.590576] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.599234] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6c5786-f978-45e5-96a8-5a0ceb0cab13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.608587] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.638042] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5beeaeb7-e79c-4ff8-a809-05cfc81d6827 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.646382] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0facd973-6fa1-497d-a12a-f846f237754a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.661459] env[65121]: DEBUG nova.compute.provider_tree [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.723533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Releasing lock "refresh_cache-9a248b47-a34b-4a98-baf0-aa971b5aca94" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.723835] env[65121]: DEBUG nova.objects.instance [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lazy-loading 'migration_context' on Instance uuid 9a248b47-a34b-4a98-baf0-aa971b5aca94 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.730732] env[65121]: DEBUG nova.network.neutron [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Successfully created port: 747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1187.838815] env[65121]: DEBUG nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1188.102554] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107452, 'name': PowerOffVM_Task, 'duration_secs': 0.406207} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.102824] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1188.103019] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1188.103209] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993563', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'name': 'volume-34ae2906-5e27-4132-9339-200c8a9a6ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '84ae6061-a962-4a74-8661-9718cc4c5346', 'attached_at': '2025-12-12T14:30:06.000000', 'detached_at': '', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'serial': '34ae2906-5e27-4132-9339-200c8a9a6ead'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1188.104038] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c682cbf-a5fb-4f71-a5d0-8afeb17fe027 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.127881] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd610b3-c8b0-47cd-9eff-59c1e2d6ce04 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.137821] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c658d57b-87cc-4081-bc92-99b35b3c1a2a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.160492] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cc0fb1-1faa-4337-88e4-65a83b91bae6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.164345] env[65121]: DEBUG nova.scheduler.client.report [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1188.178574] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The volume has not been displaced from its original location: [datastore2] volume-34ae2906-5e27-4132-9339-200c8a9a6ead/volume-34ae2906-5e27-4132-9339-200c8a9a6ead.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1188.183944] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1188.185122] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d2ce2f6-0bd5-40ae-b76c-7513fb4a0f8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.205051] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1188.205051] env[65121]: value = "task-5107453" [ 1188.205051] env[65121]: _type = "Task" [ 1188.205051] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.215689] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107453, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.228715] env[65121]: DEBUG nova.objects.base [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Object Instance<9a248b47-a34b-4a98-baf0-aa971b5aca94> lazy-loaded attributes: info_cache,migration_context {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1188.229884] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df58020-542d-4919-9ac3-3067e803f3c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.249761] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-321149d2-5cf7-4251-9d66-27066ebc55af {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.255852] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1188.255852] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b71f6-0960-8727-e703-5ad70b7f8dd1" [ 1188.255852] env[65121]: _type = "Task" [ 1188.255852] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.264210] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524b71f6-0960-8727-e703-5ad70b7f8dd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.552586] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.553328] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c380edb-8dfc-489b-9038-1a38f2e644ec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.562605] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1188.562605] env[65121]: value = "task-5107454" [ 1188.562605] env[65121]: _type = "Task" [ 1188.562605] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.572904] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.686200] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.690184] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.334s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.692284] env[65121]: INFO nova.compute.claims [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1188.716512] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107453, 'name': ReconfigVM_Task, 'duration_secs': 0.454776} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.717691] env[65121]: INFO nova.scheduler.client.report [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Deleted allocations for instance 3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7 [ 1188.719409] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1188.728089] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c23fd41d-3305-47bc-88bc-155f5ab7b696 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.744697] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1188.744697] env[65121]: value = "task-5107455" [ 1188.744697] env[65121]: _type = "Task" [ 1188.744697] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.756135] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107455, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.768497] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524b71f6-0960-8727-e703-5ad70b7f8dd1, 'name': SearchDatastore_Task, 'duration_secs': 0.009766} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.768839] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.851036] env[65121]: DEBUG nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1188.883798] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1188.884077] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1188.884228] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1188.884412] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1188.884551] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1188.884687] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1188.884882] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1188.885038] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1188.885209] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1188.885362] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1188.885520] env[65121]: DEBUG nova.virt.hardware [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1188.886932] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4630e81d-0580-45b7-9cf2-5e8011e2715d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.896010] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b1dc58-08ba-454e-8459-ec5e20dfeb73 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.073275] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107454, 'name': PowerOffVM_Task, 'duration_secs': 0.187774} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.073556] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1189.074406] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a936c0-d390-433a-9565-fa7e4f5c28c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.094539] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bd4fcb-5b3f-4336-b6c7-8e195893cb5f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.129336] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1189.129662] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64f9bc99-6750-4a48-8177-b539842d9d1a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.137287] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1189.137287] env[65121]: value = "task-5107456" [ 1189.137287] env[65121]: _type = "Task" [ 1189.137287] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.146630] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1189.146792] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1189.147134] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.147291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.147473] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1189.147731] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0d8e22d-fd72-404c-8fd1-d9517480679f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.160575] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1189.160769] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1189.161618] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-108f14ec-c6f3-45c5-bbda-1f94d550ccc4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.167919] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1189.167919] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524e6073-8473-2306-6b29-9e21e7ce48d1" [ 1189.167919] env[65121]: _type = "Task" [ 1189.167919] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.176601] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524e6073-8473-2306-6b29-9e21e7ce48d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.231172] env[65121]: DEBUG oslo_concurrency.lockutils [None req-bcb97596-69f1-4947-a103-7071718e0bfd tempest-AttachInterfacesTestJSON-765137444 tempest-AttachInterfacesTestJSON-765137444-project-member] Lock "3ea0948e-f605-4fe5-bcd1-fbc2159ec2d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.030s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.257873] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107455, 'name': ReconfigVM_Task, 'duration_secs': 0.164047} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.258215] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993563', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'name': 'volume-34ae2906-5e27-4132-9339-200c8a9a6ead', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '84ae6061-a962-4a74-8661-9718cc4c5346', 'attached_at': '2025-12-12T14:30:06.000000', 'detached_at': '', 'volume_id': '34ae2906-5e27-4132-9339-200c8a9a6ead', 'serial': '34ae2906-5e27-4132-9339-200c8a9a6ead'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1189.258650] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1189.259522] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b12b050e-118a-476f-a92c-8ba541fd1bcd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.268257] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1189.268535] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bd62b95-1004-44a6-825a-3abab8aa0d36 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.302589] env[65121]: DEBUG nova.network.neutron [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Successfully updated port: 747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1189.345364] env[65121]: DEBUG nova.compute.manager [req-7c365416-a7b0-448b-8d93-f851bdb60eb6 req-d0e59981-91f2-439f-8719-43e7a8adbf03 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Received event network-vif-plugged-747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1189.345364] env[65121]: DEBUG oslo_concurrency.lockutils [req-7c365416-a7b0-448b-8d93-f851bdb60eb6 req-d0e59981-91f2-439f-8719-43e7a8adbf03 service nova] Acquiring lock "992ba328-206a-4d47-a28b-f3ef9b20817f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.345519] env[65121]: DEBUG oslo_concurrency.lockutils [req-7c365416-a7b0-448b-8d93-f851bdb60eb6 req-d0e59981-91f2-439f-8719-43e7a8adbf03 service nova] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.345766] env[65121]: DEBUG oslo_concurrency.lockutils [req-7c365416-a7b0-448b-8d93-f851bdb60eb6 req-d0e59981-91f2-439f-8719-43e7a8adbf03 service nova] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.345848] env[65121]: DEBUG nova.compute.manager [req-7c365416-a7b0-448b-8d93-f851bdb60eb6 req-d0e59981-91f2-439f-8719-43e7a8adbf03 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] No waiting events found dispatching network-vif-plugged-747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1189.345960] env[65121]: WARNING nova.compute.manager [req-7c365416-a7b0-448b-8d93-f851bdb60eb6 req-d0e59981-91f2-439f-8719-43e7a8adbf03 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Received unexpected event network-vif-plugged-747792c4-6fe8-4e3c-8829-6060b4a047f0 for instance with vm_state building and task_state spawning. [ 1189.355402] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1189.355793] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1189.355951] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleting the datastore file [datastore1] 84ae6061-a962-4a74-8661-9718cc4c5346 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1189.356361] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38e4cec3-60c4-46d2-805f-20f3e47a464e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.366065] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1189.366065] env[65121]: value = "task-5107458" [ 1189.366065] env[65121]: _type = "Task" [ 1189.366065] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.375352] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107458, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.680980] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524e6073-8473-2306-6b29-9e21e7ce48d1, 'name': SearchDatastore_Task, 'duration_secs': 0.014142} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.681778] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb22f54a-98bb-4017-b08f-5c8a5bc652b2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.689014] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1189.689014] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52e49aae-8f03-5a41-5575-054318174ee9" [ 1189.689014] env[65121]: _type = "Task" [ 1189.689014] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.702259] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e49aae-8f03-5a41-5575-054318174ee9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.806577] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.806778] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.806959] env[65121]: DEBUG nova.network.neutron [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1189.886761] env[65121]: DEBUG oslo_vmware.api [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107458, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257007} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.887566] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1189.887566] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1189.888047] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1189.888047] env[65121]: INFO nova.compute.manager [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Took 2.31 seconds to destroy the instance on the hypervisor. [ 1189.888429] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1189.892972] env[65121]: DEBUG nova.compute.manager [-] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1189.893048] env[65121]: DEBUG nova.network.neutron [-] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1189.893404] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.894389] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.894799] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.932848] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca369f04-c8c7-4c33-88ec-783f65a8ea22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.943513] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff49b198-80fd-4532-a5b0-b681a68360ba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.948116] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.981350] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd85d7c3-4f21-446d-8232-cd6abe078cbc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.990342] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df63ca78-85e2-4241-b105-9b8910a27cdd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.008582] env[65121]: DEBUG nova.compute.provider_tree [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.199962] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52e49aae-8f03-5a41-5575-054318174ee9, 'name': SearchDatastore_Task, 'duration_secs': 0.010622} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.200447] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1190.200638] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. {{(pid=65121) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1190.200912] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-924d46c8-aad5-4ca5-8a2e-3952d7f603ad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.208508] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1190.208508] env[65121]: value = "task-5107459" [ 1190.208508] env[65121]: _type = "Task" [ 1190.208508] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.218392] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107459, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.310281] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.310845] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.512450] env[65121]: DEBUG nova.scheduler.client.report [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1190.550760] env[65121]: DEBUG nova.network.neutron [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1190.629943] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.630470] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.720786] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107459, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.726325] env[65121]: WARNING neutronclient.v2_0.client [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1190.726943] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.727308] env[65121]: WARNING openstack [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.824171] env[65121]: DEBUG nova.network.neutron [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Updating instance_info_cache with network_info: [{"id": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "address": "fa:16:3e:16:1d:32", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap747792c4-6f", "ovs_interfaceid": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1190.877949] env[65121]: DEBUG nova.network.neutron [-] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1191.018285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.019067] env[65121]: DEBUG nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1191.022301] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.253s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.223910] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107459, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55651} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.224947] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. [ 1191.226076] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302756f2-c5e8-4e51-a25a-056acdbef2e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.257148] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1191.257378] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9aa14829-68ce-4816-b66d-cb6651b33e42 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.280199] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1191.280199] env[65121]: value = "task-5107460" [ 1191.280199] env[65121]: _type = "Task" [ 1191.280199] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.294042] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107460, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.328057] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1191.328442] env[65121]: DEBUG nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Instance network_info: |[{"id": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "address": "fa:16:3e:16:1d:32", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap747792c4-6f", "ovs_interfaceid": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1191.329032] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:1d:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '747792c4-6fe8-4e3c-8829-6060b4a047f0', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.338408] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1191.338408] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.338408] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-88259b9f-17d5-482e-bce3-d66ada66e2c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.363585] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.363585] env[65121]: value = "task-5107461" [ 1191.363585] env[65121]: _type = "Task" [ 1191.363585] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.373277] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107461, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.381116] env[65121]: INFO nova.compute.manager [-] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Took 1.49 seconds to deallocate network for instance. [ 1191.533023] env[65121]: DEBUG nova.compute.utils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1191.536410] env[65121]: DEBUG nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1191.536410] env[65121]: DEBUG nova.network.neutron [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1191.536410] env[65121]: WARNING neutronclient.v2_0.client [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1191.536410] env[65121]: WARNING neutronclient.v2_0.client [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1191.537888] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1191.538054] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1191.631327] env[65121]: DEBUG nova.policy [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20b286090b814f129703b45e707d9622', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a27988a62b54d349a2a8e483222701a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1191.792286] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107460, 'name': ReconfigVM_Task, 'duration_secs': 0.504471} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.793377] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1191.793614] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34a780c-a9a9-4933-b86d-00b041edb4e9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.825985] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52f91fb0-4abf-4793-b249-df15d83345c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.842639] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1191.842639] env[65121]: value = "task-5107462" [ 1191.842639] env[65121]: _type = "Task" [ 1191.842639] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.852555] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107462, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.854940] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4fc953d-3d18-4085-a8a2-e7b0d03349f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.863470] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f0088d-abc6-443d-a405-ffd5d8d4e379 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.877967] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107461, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.907487] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0294b5-2449-4e35-a516-77e9f3ed791c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.916951] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01073d0f-e8a1-4af1-a106-4b36f0c18aa7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.934700] env[65121]: DEBUG nova.compute.provider_tree [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.938682] env[65121]: DEBUG nova.compute.manager [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Received event network-changed-747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1191.939155] env[65121]: DEBUG nova.compute.manager [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Refreshing instance network info cache due to event network-changed-747792c4-6fe8-4e3c-8829-6060b4a047f0. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1191.939379] env[65121]: DEBUG oslo_concurrency.lockutils [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Acquiring lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.939379] env[65121]: DEBUG oslo_concurrency.lockutils [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Acquired lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1191.939516] env[65121]: DEBUG nova.network.neutron [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Refreshing network info cache for port 747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1191.951233] env[65121]: INFO nova.compute.manager [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Took 0.57 seconds to detach 1 volumes for instance. [ 1192.047691] env[65121]: DEBUG nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1192.107042] env[65121]: DEBUG nova.compute.manager [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1192.111777] env[65121]: INFO nova.compute.manager [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Rebuilding instance [ 1192.124287] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.125134] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.208090] env[65121]: DEBUG nova.network.neutron [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Successfully created port: c59158a9-d9f6-4465-b99a-724753944d6a {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1192.238965] env[65121]: DEBUG nova.compute.manager [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1192.240034] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c15a9b-e382-475a-9083-013febc8aef5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.354259] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107462, 'name': ReconfigVM_Task, 'duration_secs': 0.185631} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.354679] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1192.355016] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-667b0c21-92cc-4e2f-aa8b-0b063771651e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.362644] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1192.362644] env[65121]: value = "task-5107463" [ 1192.362644] env[65121]: _type = "Task" [ 1192.362644] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.373977] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.379779] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107461, 'name': CreateVM_Task, 'duration_secs': 0.538029} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.380060] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1192.380478] env[65121]: WARNING neutronclient.v2_0.client [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.380880] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.380983] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.381445] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1192.381711] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a1b6a16-0b88-451f-869a-0fb72d264bdf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.387542] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1192.387542] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5290d969-0c39-0349-b2ac-fe101e0932b7" [ 1192.387542] env[65121]: _type = "Task" [ 1192.387542] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.398336] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5290d969-0c39-0349-b2ac-fe101e0932b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.443290] env[65121]: DEBUG nova.scheduler.client.report [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1192.448782] env[65121]: WARNING neutronclient.v2_0.client [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.449723] env[65121]: WARNING openstack [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.450256] env[65121]: WARNING openstack [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.460021] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1192.601596] env[65121]: WARNING openstack [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.602039] env[65121]: WARNING openstack [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.626715] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1192.720085] env[65121]: WARNING neutronclient.v2_0.client [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.720776] env[65121]: WARNING openstack [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.721296] env[65121]: WARNING openstack [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.810045] env[65121]: DEBUG nova.network.neutron [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Updated VIF entry in instance network info cache for port 747792c4-6fe8-4e3c-8829-6060b4a047f0. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1192.810577] env[65121]: DEBUG nova.network.neutron [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Updating instance_info_cache with network_info: [{"id": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "address": "fa:16:3e:16:1d:32", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap747792c4-6f", "ovs_interfaceid": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1192.874127] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107463, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.899780] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5290d969-0c39-0349-b2ac-fe101e0932b7, 'name': SearchDatastore_Task, 'duration_secs': 0.012073} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.900113] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.900351] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.900583] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.900718] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.900891] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.901193] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-faad2f0a-1b1d-4d1c-b0d5-18d789153cec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.913138] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.913328] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1192.914145] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d552ebc9-0fe1-4930-9fb4-9410df052388 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.920057] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1192.920057] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5208f012-28a5-82e2-fe40-36814b7a76e5" [ 1192.920057] env[65121]: _type = "Task" [ 1192.920057] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.928380] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5208f012-28a5-82e2-fe40-36814b7a76e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.060397] env[65121]: DEBUG nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1193.089584] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1193.089826] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1193.090170] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1193.090414] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1193.090598] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1193.090750] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1193.090957] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1193.091126] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1193.091299] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1193.091453] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1193.091617] env[65121]: DEBUG nova.virt.hardware [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1193.092562] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379feed4-7196-44ad-b3d2-2c4145b50c37 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.101991] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5915f621-52e1-4a3b-b27a-811821ed27c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.254914] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1193.256615] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2e48953-7165-476b-9545-9646f19134ff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.266605] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1193.266605] env[65121]: value = "task-5107464" [ 1193.266605] env[65121]: _type = "Task" [ 1193.266605] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.278190] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107464, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.314157] env[65121]: DEBUG oslo_concurrency.lockutils [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] Releasing lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.314510] env[65121]: DEBUG nova.compute.manager [req-81222588-9a47-40f2-b1dc-574bcf4521f3 req-33fa61dd-b5c1-4c30-a346-8f5289de67f2 service nova] [instance: 84ae6061-a962-4a74-8661-9718cc4c5346] Received event network-vif-deleted-2acd5dbc-a08c-4ebd-922b-284294369a33 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1193.377967] env[65121]: DEBUG oslo_vmware.api [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107463, 'name': PowerOnVM_Task, 'duration_secs': 0.554864} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.378292] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1193.381501] env[65121]: DEBUG nova.compute.manager [None req-1c5e4c93-a789-4ae8-920d-3be99bb2e6b6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1193.382347] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec6bcf2-50c0-4f44-8b4f-77e70130a505 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.432826] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5208f012-28a5-82e2-fe40-36814b7a76e5, 'name': SearchDatastore_Task, 'duration_secs': 0.021439} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.434107] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78099692-e235-47db-ad10-14b3f440ecad {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.443146] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1193.443146] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5238cd8e-d1be-31c2-e15c-4b39ced090be" [ 1193.443146] env[65121]: _type = "Task" [ 1193.443146] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.457209] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5238cd8e-d1be-31c2-e15c-4b39ced090be, 'name': SearchDatastore_Task, 'duration_secs': 0.011368} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.457586] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.458018] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 992ba328-206a-4d47-a28b-f3ef9b20817f/992ba328-206a-4d47-a28b-f3ef9b20817f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1193.458396] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bc6ac02-0116-487d-8749-8e084e0a8a96 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.465255] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.443s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.469884] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.010s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.470126] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.472357] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.846s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.473945] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1193.473945] env[65121]: value = "task-5107465" [ 1193.473945] env[65121]: _type = "Task" [ 1193.473945] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.484128] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.499290] env[65121]: INFO nova.scheduler.client.report [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted allocations for instance 84ae6061-a962-4a74-8661-9718cc4c5346 [ 1193.779470] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107464, 'name': PowerOffVM_Task, 'duration_secs': 0.298641} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.782330] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1193.783431] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1193.783936] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09d0aa32-f363-4dff-b70c-a2414bfbdeb3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.794779] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1193.794779] env[65121]: value = "task-5107466" [ 1193.794779] env[65121]: _type = "Task" [ 1193.794779] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.809335] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1193.810348] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1193.810501] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993557', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'name': 'volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4f86672c-425d-4641-a68c-2d6f5324e382', 'attached_at': '', 'detached_at': '', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'serial': '20d5c722-822e-46f0-9b9b-7f7d92bd114a'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1193.811461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48187e3a-aad5-40f4-9e68-9219d9ced8dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.841893] env[65121]: DEBUG nova.network.neutron [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Successfully updated port: c59158a9-d9f6-4465-b99a-724753944d6a {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1193.845273] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ea0fa3-1bb9-4799-8fa2-e07282920d7e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.858319] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f3234c-cd0c-48d2-87a2-407df072010a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.883706] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303f60cb-640d-4a5c-9c97-c251108bf6d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.907129] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] The volume has not been displaced from its original location: [datastore1] volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a/volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1193.912902] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Reconfiguring VM instance instance-0000006d to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1193.914543] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47899a3c-3e07-46f8-a6b9-bd62d4c9eefb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.938467] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1193.938467] env[65121]: value = "task-5107467" [ 1193.938467] env[65121]: _type = "Task" [ 1193.938467] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.959991] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107467, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.981048] env[65121]: INFO nova.compute.claims [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1193.999084] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505986} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.999409] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 992ba328-206a-4d47-a28b-f3ef9b20817f/992ba328-206a-4d47-a28b-f3ef9b20817f.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1193.999646] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1193.999963] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82b91cee-207e-4756-b2f7-a8ac9c4e4813 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.007982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-80168e82-1431-4bc0-ae7f-6dd1393b5941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "84ae6061-a962-4a74-8661-9718cc4c5346" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.935s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.010957] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1194.010957] env[65121]: value = "task-5107468" [ 1194.010957] env[65121]: _type = "Task" [ 1194.010957] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.026744] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.059969] env[65121]: INFO nova.scheduler.client.report [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted allocation for migration 94b4d03a-9745-4cc1-bed1-3ff1d03f3bfb [ 1194.262686] env[65121]: DEBUG nova.compute.manager [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Received event network-vif-plugged-c59158a9-d9f6-4465-b99a-724753944d6a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1194.262940] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Acquiring lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1194.263183] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1194.263550] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.263550] env[65121]: DEBUG nova.compute.manager [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] No waiting events found dispatching network-vif-plugged-c59158a9-d9f6-4465-b99a-724753944d6a {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1194.263747] env[65121]: WARNING nova.compute.manager [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Received unexpected event network-vif-plugged-c59158a9-d9f6-4465-b99a-724753944d6a for instance with vm_state building and task_state spawning. [ 1194.263830] env[65121]: DEBUG nova.compute.manager [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Received event network-changed-c59158a9-d9f6-4465-b99a-724753944d6a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1194.263995] env[65121]: DEBUG nova.compute.manager [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Refreshing instance network info cache due to event network-changed-c59158a9-d9f6-4465-b99a-724753944d6a. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1194.264214] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Acquiring lock "refresh_cache-fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.264293] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Acquired lock "refresh_cache-fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1194.264435] env[65121]: DEBUG nova.network.neutron [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Refreshing network info cache for port c59158a9-d9f6-4465-b99a-724753944d6a {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1194.350084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "refresh_cache-fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.453110] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107467, 'name': ReconfigVM_Task, 'duration_secs': 0.27149} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.453535] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Reconfigured VM instance instance-0000006d to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1194.458614] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4adf714b-b4af-41b7-8b4f-3c8c4323e101 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.475839] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1194.475839] env[65121]: value = "task-5107469" [ 1194.475839] env[65121]: _type = "Task" [ 1194.475839] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.486047] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107469, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.488105] env[65121]: INFO nova.compute.resource_tracker [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating resource usage from migration b70e77be-ee4c-45d8-9e0a-65a78916fa41 [ 1194.527120] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115245} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.531832] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1194.533607] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a08cb5-0552-44c1-bd5e-8864c354b0a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.566601] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] 992ba328-206a-4d47-a28b-f3ef9b20817f/992ba328-206a-4d47-a28b-f3ef9b20817f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1194.570603] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.230s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.571633] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1716fef6-f15c-4ddf-8f98-90359d1eed7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.595530] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1194.595530] env[65121]: value = "task-5107470" [ 1194.595530] env[65121]: _type = "Task" [ 1194.595530] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.608146] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107470, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.732116] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7a82d4-55a9-4c61-a3cc-0a68c11e6e2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.743486] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b754fd2-0f52-4033-8060-9dcf3e6ed58e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.784827] env[65121]: WARNING neutronclient.v2_0.client [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1194.785734] env[65121]: WARNING openstack [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1194.786099] env[65121]: WARNING openstack [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.796040] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0fc6d6-7040-42b1-8768-ac36ebe9ae64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.805258] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c8f1da-d4b1-4fb4-bfaa-e406309eeab6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.821935] env[65121]: DEBUG nova.compute.provider_tree [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.831572] env[65121]: DEBUG nova.network.neutron [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1194.953584] env[65121]: DEBUG nova.network.neutron [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.986791] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107469, 'name': ReconfigVM_Task, 'duration_secs': 0.326981} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.987058] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993557', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'name': 'volume-20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '4f86672c-425d-4641-a68c-2d6f5324e382', 'attached_at': '', 'detached_at': '', 'volume_id': '20d5c722-822e-46f0-9b9b-7f7d92bd114a', 'serial': '20d5c722-822e-46f0-9b9b-7f7d92bd114a'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1194.987405] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1194.988265] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbb8064-cde0-4179-83fe-4086f7dbd87c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.996228] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1194.996703] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfadc623-fd3e-479c-a990-54ad28ffbfa1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.074019] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1195.074265] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1195.074637] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Deleting the datastore file [datastore1] 4f86672c-425d-4641-a68c-2d6f5324e382 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1195.074768] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-514316d7-ebcc-492f-be63-25fa82831868 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.083775] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for the task: (returnval){ [ 1195.083775] env[65121]: value = "task-5107472" [ 1195.083775] env[65121]: _type = "Task" [ 1195.083775] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.095155] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.105432] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107470, 'name': ReconfigVM_Task, 'duration_secs': 0.410563} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.105819] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Reconfigured VM instance instance-00000076 to attach disk [datastore1] 992ba328-206a-4d47-a28b-f3ef9b20817f/992ba328-206a-4d47-a28b-f3ef9b20817f.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1195.106625] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ac07610-93e8-494d-bd37-a8fa58321869 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.115523] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1195.115523] env[65121]: value = "task-5107473" [ 1195.115523] env[65121]: _type = "Task" [ 1195.115523] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.126437] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107473, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.325521] env[65121]: DEBUG nova.scheduler.client.report [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1195.456160] env[65121]: DEBUG oslo_concurrency.lockutils [req-1f7093b3-a129-4648-985a-d93e092f9d03 req-404b901f-85cb-47aa-bac1-27ae72c863d4 service nova] Releasing lock "refresh_cache-fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1195.457578] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquired lock "refresh_cache-fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1195.457978] env[65121]: DEBUG nova.network.neutron [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1195.596188] env[65121]: DEBUG oslo_vmware.api [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Task: {'id': task-5107472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094891} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.597922] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1195.598144] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1195.599236] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1195.628011] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107473, 'name': Rename_Task, 'duration_secs': 0.162978} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.628334] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1195.628611] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4645ca06-9dfa-43f5-ad18-9d99dd9ead76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.638176] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1195.638176] env[65121]: value = "task-5107474" [ 1195.638176] env[65121]: _type = "Task" [ 1195.638176] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.647873] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.677150] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1195.677150] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a0416e4-4664-4571-962b-8cc518dc3e01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.687691] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14513bd-49dd-49c7-9143-c2882254f5ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.737896] env[65121]: ERROR nova.compute.manager [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Failed to detach volume 20d5c722-822e-46f0-9b9b-7f7d92bd114a from /dev/sda: nova.exception.InstanceNotFound: Instance 4f86672c-425d-4641-a68c-2d6f5324e382 could not be found. [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Traceback (most recent call last): [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self.driver.rebuild(**kwargs) [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] raise NotImplementedError() [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] NotImplementedError [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] During handling of the above exception, another exception occurred: [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Traceback (most recent call last): [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self.driver.detach_volume(context, old_connection_info, [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] return self._volumeops.detach_volume(connection_info, instance) [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self._detach_volume_vmdk(connection_info, instance) [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] stable_ref.fetch_moref(session) [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] nova.exception.InstanceNotFound: Instance 4f86672c-425d-4641-a68c-2d6f5324e382 could not be found. [ 1195.737896] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.831656] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.359s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.831957] env[65121]: INFO nova.compute.manager [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Migrating [ 1195.860143] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.860143] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.870581] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.873094] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.873094] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.873094] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.873094] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.880448] env[65121]: INFO nova.compute.manager [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Terminating instance [ 1195.927342] env[65121]: DEBUG nova.compute.utils [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Build of instance 4f86672c-425d-4641-a68c-2d6f5324e382 aborted: Failed to rebuild volume backed instance. {{(pid=65121) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1195.932070] env[65121]: ERROR nova.compute.manager [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 4f86672c-425d-4641-a68c-2d6f5324e382 aborted: Failed to rebuild volume backed instance. [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Traceback (most recent call last): [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self.driver.rebuild(**kwargs) [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] raise NotImplementedError() [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] NotImplementedError [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] During handling of the above exception, another exception occurred: [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Traceback (most recent call last): [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 3699, in _rebuild_volume_backed_instance [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self._detach_root_volume(context, instance, root_bdm) [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 3678, in _detach_root_volume [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] with excutils.save_and_reraise_exception(): [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self.force_reraise() [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] raise self.value [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self.driver.detach_volume(context, old_connection_info, [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] return self._volumeops.detach_volume(connection_info, instance) [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self._detach_volume_vmdk(connection_info, instance) [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] stable_ref.fetch_moref(session) [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] nova.exception.InstanceNotFound: Instance 4f86672c-425d-4641-a68c-2d6f5324e382 could not be found. [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] During handling of the above exception, another exception occurred: [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Traceback (most recent call last): [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 11570, in _error_out_instance_on_exception [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] yield [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 3967, in rebuild_instance [ 1195.932070] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self._do_rebuild_instance_with_claim( [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 4053, in _do_rebuild_instance_with_claim [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self._do_rebuild_instance( [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 4245, in _do_rebuild_instance [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self._rebuild_default_impl(**kwargs) [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 3822, in _rebuild_default_impl [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] self._rebuild_volume_backed_instance( [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] File "/opt/stack/nova/nova/compute/manager.py", line 3714, in _rebuild_volume_backed_instance [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] raise exception.BuildAbortException( [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] nova.exception.BuildAbortException: Build of instance 4f86672c-425d-4641-a68c-2d6f5324e382 aborted: Failed to rebuild volume backed instance. [ 1195.933351] env[65121]: ERROR nova.compute.manager [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] [ 1195.961635] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.962080] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.008433] env[65121]: DEBUG nova.network.neutron [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1196.037151] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.037519] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.112379] env[65121]: WARNING neutronclient.v2_0.client [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.113291] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.113589] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.153313] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107474, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.163930] env[65121]: INFO nova.compute.manager [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Unrescuing [ 1196.164611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.164611] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.164611] env[65121]: DEBUG nova.network.neutron [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1196.213507] env[65121]: DEBUG nova.network.neutron [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Updating instance_info_cache with network_info: [{"id": "c59158a9-d9f6-4465-b99a-724753944d6a", "address": "fa:16:3e:2f:3b:9a", "network": {"id": "0b61182e-a953-4bd2-a909-3ac9c3e5ba71", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1542431405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a27988a62b54d349a2a8e483222701a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc59158a9-d9", "ovs_interfaceid": "c59158a9-d9f6-4465-b99a-724753944d6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1196.351170] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.351614] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.352382] env[65121]: DEBUG nova.network.neutron [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1196.364077] env[65121]: DEBUG nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1196.385961] env[65121]: DEBUG nova.compute.manager [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1196.386088] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1196.387519] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8145e47-12b6-41f0-afef-9e5b127f84b4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.397034] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.397034] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-158cc97f-6e8a-4da7-ac54-9e18624e5933 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.404640] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1196.404640] env[65121]: value = "task-5107475" [ 1196.404640] env[65121]: _type = "Task" [ 1196.404640] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.415274] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.655264] env[65121]: DEBUG oslo_vmware.api [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107474, 'name': PowerOnVM_Task, 'duration_secs': 0.598466} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.655264] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1196.655264] env[65121]: INFO nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Took 7.80 seconds to spawn the instance on the hypervisor. [ 1196.655264] env[65121]: DEBUG nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1196.655264] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a09558-6c37-4cfb-8f55-3692cbd8f0bf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.669437] env[65121]: WARNING neutronclient.v2_0.client [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.669437] env[65121]: WARNING openstack [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.669437] env[65121]: WARNING openstack [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.718627] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Releasing lock "refresh_cache-fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.718627] env[65121]: DEBUG nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Instance network_info: |[{"id": "c59158a9-d9f6-4465-b99a-724753944d6a", "address": "fa:16:3e:2f:3b:9a", "network": {"id": "0b61182e-a953-4bd2-a909-3ac9c3e5ba71", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1542431405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a27988a62b54d349a2a8e483222701a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc59158a9-d9", "ovs_interfaceid": "c59158a9-d9f6-4465-b99a-724753944d6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1196.718627] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:3b:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c59158a9-d9f6-4465-b99a-724753944d6a', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1196.726179] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Creating folder: Project (8a27988a62b54d349a2a8e483222701a). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1196.727961] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a6ad66e-c554-450a-83be-f886a2ba6da7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.743312] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Created folder: Project (8a27988a62b54d349a2a8e483222701a) in parent group-v993268. [ 1196.744356] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Creating folder: Instances. Parent ref: group-v993583. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1196.744356] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81a8d256-c06c-4fa4-a574-9c3821bbf32f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.754898] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Created folder: Instances in parent group-v993583. [ 1196.755814] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1196.756073] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1196.756307] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32ded477-d000-4a04-80b2-b2c97d6ac875 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.779280] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1196.779280] env[65121]: value = "task-5107478" [ 1196.779280] env[65121]: _type = "Task" [ 1196.779280] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.788715] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107478, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.845096] env[65121]: WARNING openstack [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.845631] env[65121]: WARNING openstack [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.855120] env[65121]: WARNING neutronclient.v2_0.client [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.855768] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.856118] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.898803] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1196.899106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1196.900731] env[65121]: INFO nova.compute.claims [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1196.915842] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107475, 'name': PowerOffVM_Task, 'duration_secs': 0.234279} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.916158] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.916328] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.916738] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e6a4cd2-895b-41ab-a604-1da483487aed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.943462] env[65121]: WARNING neutronclient.v2_0.client [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.944140] env[65121]: WARNING openstack [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.944478] env[65121]: WARNING openstack [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.987084] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.987430] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.987717] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleting the datastore file [datastore1] 9a248b47-a34b-4a98-baf0-aa971b5aca94 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.988128] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1493fb09-0171-4c00-bff8-bcf13e174eeb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.997356] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for the task: (returnval){ [ 1196.997356] env[65121]: value = "task-5107480" [ 1196.997356] env[65121]: _type = "Task" [ 1196.997356] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.009232] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.075955] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.076438] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.112998] env[65121]: DEBUG nova.network.neutron [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Updating instance_info_cache with network_info: [{"id": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "address": "fa:16:3e:19:1a:cf", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf714bfb3-87", "ovs_interfaceid": "f714bfb3-87de-4c6b-ba80-8227e6eebf65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.173316] env[65121]: INFO nova.compute.manager [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Took 13.00 seconds to build instance. [ 1197.193775] env[65121]: WARNING neutronclient.v2_0.client [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.194632] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.195062] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.292747] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107478, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.302124] env[65121]: DEBUG nova.network.neutron [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.512152] env[65121]: DEBUG oslo_vmware.api [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Task: {'id': task-5107480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247823} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.512152] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1197.512449] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1197.512562] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1197.514162] env[65121]: INFO nova.compute.manager [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1197.514484] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1197.515383] env[65121]: DEBUG nova.compute.manager [-] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1197.515496] env[65121]: DEBUG nova.network.neutron [-] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1197.515756] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.516301] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.516586] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.591236] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.619266] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-3859dd02-9b52-4c71-a41f-c41f47ac1706" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.620030] env[65121]: DEBUG nova.objects.instance [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lazy-loading 'flavor' on Instance uuid 3859dd02-9b52-4c71-a41f-c41f47ac1706 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.677084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-88285732-d8e2-4ec8-9230-870a3d35a151 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.514s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.792189] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107478, 'name': CreateVM_Task, 'duration_secs': 0.591051} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.792189] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1197.792189] env[65121]: WARNING neutronclient.v2_0.client [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.792707] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.793077] env[65121]: WARNING openstack [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.806124] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.886919] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.887112] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1197.887449] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1197.887779] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a84593ca-6c81-4267-b39c-b73e47732e3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.896538] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1197.896538] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5237abd5-0857-ba36-f756-b23f37745016" [ 1197.896538] env[65121]: _type = "Task" [ 1197.896538] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.907451] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5237abd5-0857-ba36-f756-b23f37745016, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.961657] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1197.978023] env[65121]: DEBUG nova.compute.manager [req-365ad2ff-ffad-437c-948f-08b218bd054c req-80049298-cb2a-474b-a6da-122b0215ce5c service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Received event network-vif-deleted-d86b7885-6e5e-45ec-b64e-19ae24268b82 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1197.978425] env[65121]: INFO nova.compute.manager [req-365ad2ff-ffad-437c-948f-08b218bd054c req-80049298-cb2a-474b-a6da-122b0215ce5c service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Neutron deleted interface d86b7885-6e5e-45ec-b64e-19ae24268b82; detaching it from the instance and deleting it from the info cache [ 1197.978425] env[65121]: DEBUG nova.network.neutron [req-365ad2ff-ffad-437c-948f-08b218bd054c req-80049298-cb2a-474b-a6da-122b0215ce5c service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1198.126494] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455f9f10-2d42-4ef4-95cd-8d59c3a2750c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.150869] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1198.154036] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-622e14fe-1475-406f-a007-c33d7a6f680c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.162631] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "e19d6d9b-d15d-4780-8564-f479573ce59b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.162847] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.164333] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1198.164333] env[65121]: value = "task-5107481" [ 1198.164333] env[65121]: _type = "Task" [ 1198.164333] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.170401] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877a4f95-17bf-4831-8198-f08af1b2e07c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.181331] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107481, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.183206] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032dc960-569a-454d-91be-c7fec17d8123 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.218125] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e61848e-f502-490f-bcfe-f32a95230582 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.226851] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2971844a-26d9-418c-b030-68f3a9180f16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.245452] env[65121]: DEBUG nova.compute.provider_tree [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.393016] env[65121]: DEBUG nova.network.neutron [-] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1198.410696] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5237abd5-0857-ba36-f756-b23f37745016, 'name': SearchDatastore_Task, 'duration_secs': 0.013663} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.413667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1198.413667] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1198.413667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.413667] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1198.413667] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1198.413667] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d693fd0-3a43-4e3f-8264-b35487e0e8fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.426666] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1198.426666] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1198.426666] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f1d4d91-25de-446c-bb32-847eb90a2994 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.434500] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1198.434500] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]528a0c6d-e264-d8fc-2d69-33fbf5a181e6" [ 1198.434500] env[65121]: _type = "Task" [ 1198.434500] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.448319] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528a0c6d-e264-d8fc-2d69-33fbf5a181e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.481732] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5de0598-656d-4200-b4f3-f1175adceb40 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.493404] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d658c0-bde0-4173-b23b-965ced34af13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.508610] env[65121]: DEBUG nova.compute.manager [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Received event network-changed-747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1198.508843] env[65121]: DEBUG nova.compute.manager [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Refreshing instance network info cache due to event network-changed-747792c4-6fe8-4e3c-8829-6060b4a047f0. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1198.509122] env[65121]: DEBUG oslo_concurrency.lockutils [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Acquiring lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.509305] env[65121]: DEBUG oslo_concurrency.lockutils [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Acquired lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1198.509523] env[65121]: DEBUG nova.network.neutron [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Refreshing network info cache for port 747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1198.538072] env[65121]: DEBUG nova.compute.manager [req-365ad2ff-ffad-437c-948f-08b218bd054c req-80049298-cb2a-474b-a6da-122b0215ce5c service nova] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Detach interface failed, port_id=d86b7885-6e5e-45ec-b64e-19ae24268b82, reason: Instance 9a248b47-a34b-4a98-baf0-aa971b5aca94 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1198.568353] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquiring lock "4f86672c-425d-4641-a68c-2d6f5324e382" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.568704] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "4f86672c-425d-4641-a68c-2d6f5324e382" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.569095] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquiring lock "4f86672c-425d-4641-a68c-2d6f5324e382-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.569239] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "4f86672c-425d-4641-a68c-2d6f5324e382-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.569302] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "4f86672c-425d-4641-a68c-2d6f5324e382-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1198.571778] env[65121]: INFO nova.compute.manager [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Terminating instance [ 1198.670455] env[65121]: DEBUG nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1198.683028] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107481, 'name': PowerOffVM_Task, 'duration_secs': 0.230145} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.683368] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1198.688968] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Reconfiguring VM instance instance-00000071 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1198.691658] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1dfb8b9c-4354-42c9-95de-94dc29000e41 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.715099] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1198.715099] env[65121]: value = "task-5107482" [ 1198.715099] env[65121]: _type = "Task" [ 1198.715099] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.727809] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107482, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.750293] env[65121]: DEBUG nova.scheduler.client.report [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1198.897191] env[65121]: INFO nova.compute.manager [-] [instance: 9a248b47-a34b-4a98-baf0-aa971b5aca94] Took 1.38 seconds to deallocate network for instance. [ 1198.950453] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]528a0c6d-e264-d8fc-2d69-33fbf5a181e6, 'name': SearchDatastore_Task, 'duration_secs': 0.012532} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.951457] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ab34570-21e1-4c77-ac1f-6e3d2d8a2faa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.959265] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1198.959265] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52110caa-0a31-bdf8-a244-7b3569a562f3" [ 1198.959265] env[65121]: _type = "Task" [ 1198.959265] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.969207] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52110caa-0a31-bdf8-a244-7b3569a562f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.014431] env[65121]: WARNING neutronclient.v2_0.client [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1199.015201] env[65121]: WARNING openstack [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1199.015477] env[65121]: WARNING openstack [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1199.076029] env[65121]: DEBUG nova.compute.manager [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1199.076029] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ad7114b-8ba3-45e0-be99-59f3006f45a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.089067] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c99c86-5baf-4cc0-8831-8b729ab1050d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.123822] env[65121]: WARNING nova.virt.vmwareapi.driver [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 4f86672c-425d-4641-a68c-2d6f5324e382 could not be found. [ 1199.123964] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1199.128864] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecb9904c-6a19-41bf-88d1-49728ff187d7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.138837] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cb144a-1e47-4c2b-97ef-97a997f46866 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.150757] env[65121]: WARNING openstack [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1199.151135] env[65121]: WARNING openstack [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1199.187268] env[65121]: WARNING nova.virt.vmwareapi.vmops [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f86672c-425d-4641-a68c-2d6f5324e382 could not be found. [ 1199.187469] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1199.187642] env[65121]: INFO nova.compute.manager [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Took 0.11 seconds to destroy the instance on the hypervisor. [ 1199.187884] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1199.194560] env[65121]: DEBUG nova.compute.manager [-] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1199.194665] env[65121]: DEBUG nova.network.neutron [-] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1199.194887] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1199.195408] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1199.195773] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1199.221208] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.227415] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107482, 'name': ReconfigVM_Task, 'duration_secs': 0.24112} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.227756] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Reconfigured VM instance instance-00000071 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1199.227927] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1199.228226] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ee90b7d-3a71-4e91-852e-0211fc1b7d0a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.235538] env[65121]: WARNING neutronclient.v2_0.client [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1199.236312] env[65121]: WARNING openstack [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1199.236709] env[65121]: WARNING openstack [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1199.246475] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1199.246475] env[65121]: value = "task-5107483" [ 1199.246475] env[65121]: _type = "Task" [ 1199.246475] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.256426] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1199.257049] env[65121]: DEBUG nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1199.260033] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107483, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.261113] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1199.263298] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.302s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.322937] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78c9ae4-fe67-4339-b5dd-faedfadd26a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.345170] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance 'f58bf796-69e7-4a61-9cea-78dff8d4fab5' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1199.353734] env[65121]: DEBUG nova.network.neutron [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Updated VIF entry in instance network info cache for port 747792c4-6fe8-4e3c-8829-6060b4a047f0. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1199.354131] env[65121]: DEBUG nova.network.neutron [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Updating instance_info_cache with network_info: [{"id": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "address": "fa:16:3e:16:1d:32", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap747792c4-6f", "ovs_interfaceid": "747792c4-6fe8-4e3c-8829-6060b4a047f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1199.404718] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.472323] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52110caa-0a31-bdf8-a244-7b3569a562f3, 'name': SearchDatastore_Task, 'duration_secs': 0.012416} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.475129] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1199.475422] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] fb69eaef-32e9-4ade-9ed2-27ed1df36ce3/fb69eaef-32e9-4ade-9ed2-27ed1df36ce3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1199.475897] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd4522ee-387f-401f-9e2e-ffe6cabe6bdb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.484052] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1199.484052] env[65121]: value = "task-5107484" [ 1199.484052] env[65121]: _type = "Task" [ 1199.484052] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.494496] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107484, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.512345] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ffba03-f2df-4c8d-bb5c-ad7a0ddc561b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.521961] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49714c6d-901c-48b1-968f-641c4fe892e0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.561785] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d7acca-c81a-4b72-be78-a256c590cde1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.573035] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e8ced4-33e2-4390-b440-60741478a44c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.590827] env[65121]: DEBUG nova.compute.provider_tree [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1199.761037] env[65121]: DEBUG oslo_vmware.api [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107483, 'name': PowerOnVM_Task, 'duration_secs': 0.424163} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.761037] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1199.761037] env[65121]: DEBUG nova.compute.manager [None req-e4c2aaa2-b8fc-4129-8734-c57e0ae54eb6 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1199.761037] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b9577d-1acc-416b-9e55-38340ca581cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.767094] env[65121]: DEBUG nova.compute.utils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1199.769348] env[65121]: DEBUG nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1199.770072] env[65121]: DEBUG nova.network.neutron [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1199.770697] env[65121]: WARNING neutronclient.v2_0.client [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1199.771471] env[65121]: WARNING neutronclient.v2_0.client [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1199.772313] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1199.772847] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1199.783736] env[65121]: DEBUG nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1199.858372] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1199.859477] env[65121]: DEBUG oslo_concurrency.lockutils [req-fee57919-6013-4e66-a66b-02ed7d84a669 req-ee1cb9bd-5f7b-4b3c-bbbf-ebcc82e2d138 service nova] Releasing lock "refresh_cache-992ba328-206a-4d47-a28b-f3ef9b20817f" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1199.860582] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07eb3d48-0df9-44f4-9869-0e1c660d15f0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.873113] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1199.873113] env[65121]: value = "task-5107486" [ 1199.873113] env[65121]: _type = "Task" [ 1199.873113] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.880584] env[65121]: DEBUG nova.policy [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63bf1d576d7e44c3a8f99a75546efe52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4467da3ed41245ddbc93fc865a8b7bdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1199.893858] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107486, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.998067] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107484, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.027782] env[65121]: DEBUG nova.compute.manager [req-374635ad-7017-47c4-a5f1-23b8065fea68 req-c4a39030-aeb0-45e1-8ae1-ec5c474a666a service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Received event network-vif-deleted-1d9c1075-163b-42c1-b44d-49a4699c9930 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1200.028215] env[65121]: INFO nova.compute.manager [req-374635ad-7017-47c4-a5f1-23b8065fea68 req-c4a39030-aeb0-45e1-8ae1-ec5c474a666a service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Neutron deleted interface 1d9c1075-163b-42c1-b44d-49a4699c9930; detaching it from the instance and deleting it from the info cache [ 1200.028649] env[65121]: DEBUG nova.network.neutron [req-374635ad-7017-47c4-a5f1-23b8065fea68 req-c4a39030-aeb0-45e1-8ae1-ec5c474a666a service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1200.098679] env[65121]: DEBUG nova.scheduler.client.report [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1200.316519] env[65121]: DEBUG nova.network.neutron [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Successfully created port: 3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1200.334874] env[65121]: DEBUG nova.network.neutron [-] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1200.385995] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107486, 'name': PowerOffVM_Task, 'duration_secs': 0.247602} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.386293] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1200.386468] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance 'f58bf796-69e7-4a61-9cea-78dff8d4fab5' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1200.495530] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107484, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.666533} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.495793] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] fb69eaef-32e9-4ade-9ed2-27ed1df36ce3/fb69eaef-32e9-4ade-9ed2-27ed1df36ce3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1200.495999] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1200.496272] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0da77ce3-ec2c-411b-8516-7da3f215fcd8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.503330] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1200.503330] env[65121]: value = "task-5107487" [ 1200.503330] env[65121]: _type = "Task" [ 1200.503330] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.514798] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107487, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.532588] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a95f0a54-ee5c-4218-8394-fcbd25aa7545 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.544626] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260da540-ee0e-4c4e-894b-b58161f06cd9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.582292] env[65121]: DEBUG nova.compute.manager [req-374635ad-7017-47c4-a5f1-23b8065fea68 req-c4a39030-aeb0-45e1-8ae1-ec5c474a666a service nova] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Detach interface failed, port_id=1d9c1075-163b-42c1-b44d-49a4699c9930, reason: Instance 4f86672c-425d-4641-a68c-2d6f5324e382 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1200.606707] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.343s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.607016] env[65121]: INFO nova.compute.manager [None req-9caaaddc-32d0-4b58-8566-32af53947991 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Successfully reverted task state from rebuilding on failure for instance. [ 1200.612326] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.391s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.614297] env[65121]: INFO nova.compute.claims [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1200.795776] env[65121]: DEBUG nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1200.825139] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1200.825968] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1200.826262] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1200.826581] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1200.826845] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1200.827124] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1200.827478] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1200.827724] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1200.828041] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1200.828348] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1200.828662] env[65121]: DEBUG nova.virt.hardware [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1200.830015] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f982e243-7442-4872-a515-dc5a0b6e8d68 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.839201] env[65121]: INFO nova.compute.manager [-] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Took 1.64 seconds to deallocate network for instance. [ 1200.842959] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af014650-23b8-49dd-b9c0-70d7af897655 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.895146] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1200.895744] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1200.895956] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1200.896711] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1200.896926] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1200.897157] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1200.897441] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1200.897654] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1200.897879] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1200.898115] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1200.898354] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1200.904280] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca9a3655-ccb2-4d47-be7a-b952ebdda6ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.924586] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1200.924586] env[65121]: value = "task-5107488" [ 1200.924586] env[65121]: _type = "Task" [ 1200.924586] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.935112] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107488, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.015205] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107487, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066323} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.015500] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1201.016338] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af533292-5141-4e59-b9cc-23b311a6a3cd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.042307] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] fb69eaef-32e9-4ade-9ed2-27ed1df36ce3/fb69eaef-32e9-4ade-9ed2-27ed1df36ce3.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.042746] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75dd964a-b363-471d-b96b-91f10789fc47 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.065367] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1201.065367] env[65121]: value = "task-5107489" [ 1201.065367] env[65121]: _type = "Task" [ 1201.065367] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.075125] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107489, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.399846] env[65121]: INFO nova.compute.manager [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Took 0.56 seconds to detach 1 volumes for instance. [ 1201.402270] env[65121]: DEBUG nova.compute.manager [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] [instance: 4f86672c-425d-4641-a68c-2d6f5324e382] Deleting volume: 20d5c722-822e-46f0-9b9b-7f7d92bd114a {{(pid=65121) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 1201.437110] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107488, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.582120] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107489, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.911822] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a65aa7-256e-4a8e-9de2-8b2c7604b17a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.922424] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b399ad-8e7a-4680-ae73-98687bfd5490 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.937900] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107488, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.974265] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1201.974265] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46ae1ea-0167-434f-9e2f-516a30185aee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.986589] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60a1f94-93e0-459f-b0f0-068db3fe4ad5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.011642] env[65121]: DEBUG nova.compute.provider_tree [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.031917] env[65121]: DEBUG nova.network.neutron [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Successfully updated port: 3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1202.078535] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107489, 'name': ReconfigVM_Task, 'duration_secs': 0.814553} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.078872] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Reconfigured VM instance instance-00000077 to attach disk [datastore1] fb69eaef-32e9-4ade-9ed2-27ed1df36ce3/fb69eaef-32e9-4ade-9ed2-27ed1df36ce3.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.079607] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f5a9494-6b6a-424e-abe0-8b8e85fc4688 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.091033] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1202.091033] env[65121]: value = "task-5107491" [ 1202.091033] env[65121]: _type = "Task" [ 1202.091033] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.105486] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107491, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.141448] env[65121]: DEBUG nova.compute.manager [req-0b1162ef-7327-4f81-8be1-4a117abbea10 req-5d553a28-5b33-4b3c-8a22-0f0baa6ace9e service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-vif-plugged-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1202.141655] env[65121]: DEBUG oslo_concurrency.lockutils [req-0b1162ef-7327-4f81-8be1-4a117abbea10 req-5d553a28-5b33-4b3c-8a22-0f0baa6ace9e service nova] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.142958] env[65121]: DEBUG oslo_concurrency.lockutils [req-0b1162ef-7327-4f81-8be1-4a117abbea10 req-5d553a28-5b33-4b3c-8a22-0f0baa6ace9e service nova] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1202.143258] env[65121]: DEBUG oslo_concurrency.lockutils [req-0b1162ef-7327-4f81-8be1-4a117abbea10 req-5d553a28-5b33-4b3c-8a22-0f0baa6ace9e service nova] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.143560] env[65121]: DEBUG nova.compute.manager [req-0b1162ef-7327-4f81-8be1-4a117abbea10 req-5d553a28-5b33-4b3c-8a22-0f0baa6ace9e service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] No waiting events found dispatching network-vif-plugged-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1202.143597] env[65121]: WARNING nova.compute.manager [req-0b1162ef-7327-4f81-8be1-4a117abbea10 req-5d553a28-5b33-4b3c-8a22-0f0baa6ace9e service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received unexpected event network-vif-plugged-3cbde80d-7df5-4785-821e-b45797716ab3 for instance with vm_state building and task_state spawning. [ 1202.441724] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107488, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.516626] env[65121]: DEBUG nova.scheduler.client.report [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1202.537795] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.537983] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1202.538174] env[65121]: DEBUG nova.network.neutron [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1202.603687] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107491, 'name': Rename_Task, 'duration_secs': 0.258077} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.603955] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1202.604266] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a649bab-8eb7-4904-a1ee-1896fe56c777 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.614032] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1202.614032] env[65121]: value = "task-5107492" [ 1202.614032] env[65121]: _type = "Task" [ 1202.614032] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.624853] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107492, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.937831] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107488, 'name': ReconfigVM_Task, 'duration_secs': 1.527747} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.938186] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance 'f58bf796-69e7-4a61-9cea-78dff8d4fab5' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1203.026040] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.026382] env[65121]: DEBUG nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1203.029320] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.625s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.029540] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.033213] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.063s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.033444] env[65121]: DEBUG nova.objects.instance [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lazy-loading 'resources' on Instance uuid 4f86672c-425d-4641-a68c-2d6f5324e382 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.044286] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.044724] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.064866] env[65121]: INFO nova.scheduler.client.report [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Deleted allocations for instance 9a248b47-a34b-4a98-baf0-aa971b5aca94 [ 1203.091408] env[65121]: DEBUG nova.network.neutron [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1203.117034] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.118674] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.137525] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107492, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.249392] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1203.250472] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.251063] env[65121]: INFO nova.compute.manager [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Shelving [ 1203.260427] env[65121]: WARNING neutronclient.v2_0.client [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.261125] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.261994] env[65121]: WARNING openstack [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.392246] env[65121]: DEBUG nova.network.neutron [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": "3cbde80d-7df5-4785-821e-b45797716ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1203.445862] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1203.446206] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1203.446418] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1203.446572] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1203.446729] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1203.446853] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1203.447225] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1203.447427] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1203.447600] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1203.447761] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1203.447972] env[65121]: DEBUG nova.virt.hardware [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1203.454123] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1203.454870] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-919edde0-00bb-4173-92d3-42a1cc76a25f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.477151] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1203.477151] env[65121]: value = "task-5107494" [ 1203.477151] env[65121]: _type = "Task" [ 1203.477151] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.486861] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107494, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.537518] env[65121]: DEBUG nova.compute.utils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1203.542626] env[65121]: DEBUG nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1203.543616] env[65121]: DEBUG nova.network.neutron [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1203.543616] env[65121]: WARNING neutronclient.v2_0.client [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.543616] env[65121]: WARNING neutronclient.v2_0.client [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.545078] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.545078] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.581051] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3727fec4-138c-4756-b6cd-9b2834b21b0e tempest-DeleteServersTestJSON-1929788868 tempest-DeleteServersTestJSON-1929788868-project-member] Lock "9a248b47-a34b-4a98-baf0-aa971b5aca94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.709s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.630404] env[65121]: DEBUG nova.policy [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2471f8b61d04d4dbbc29879b01e8f8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b07f57e89c774ef1bbaff256800ab80b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1203.651185] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107492, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.811365] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f02003-5165-4bd8-a9e1-e40a82d271a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.820709] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af908a4-127b-48e6-8a87-f32f9b52787e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.856978] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d604f858-22dc-4c6b-9040-7a15f2928adf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.867028] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ed6d1e-5d65-494c-b298-74bf3f00abaf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.885157] env[65121]: DEBUG nova.compute.provider_tree [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1203.895750] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1203.896206] env[65121]: DEBUG nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Instance network_info: |[{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": "3cbde80d-7df5-4785-821e-b45797716ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1203.896924] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:17:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cbde80d-7df5-4785-821e-b45797716ab3', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1203.905564] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1203.906196] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1203.906681] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f6b1fa2-b30f-43cc-b32a-b195d5790e4b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.932866] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1203.932866] env[65121]: value = "task-5107496" [ 1203.932866] env[65121]: _type = "Task" [ 1203.932866] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.942950] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107496, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.989914] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107494, 'name': ReconfigVM_Task, 'duration_secs': 0.238109} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.990312] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1203.991461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e766ea-7786-4f8a-b29a-67dedb7f0e88 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.017197] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1204.017662] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36832679-a7b7-4e20-89cc-88c27c212f32 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.040462] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1204.040462] env[65121]: value = "task-5107497" [ 1204.040462] env[65121]: _type = "Task" [ 1204.040462] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.045202] env[65121]: DEBUG nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1204.063060] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107497, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.103312] env[65121]: DEBUG nova.network.neutron [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Successfully created port: c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1204.145633] env[65121]: DEBUG oslo_vmware.api [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107492, 'name': PowerOnVM_Task, 'duration_secs': 1.422559} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.145885] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1204.146113] env[65121]: INFO nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Took 11.09 seconds to spawn the instance on the hypervisor. [ 1204.146573] env[65121]: DEBUG nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1204.147587] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f84f31c-440f-49fb-988f-5c6ba9c18b26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.267963] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1204.268471] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c30c153a-1257-48d5-940c-997d79589303 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.273692] env[65121]: DEBUG nova.compute.manager [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1204.273966] env[65121]: DEBUG nova.compute.manager [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing instance network info cache due to event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1204.274272] env[65121]: DEBUG oslo_concurrency.lockutils [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Acquiring lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.274450] env[65121]: DEBUG oslo_concurrency.lockutils [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Acquired lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1204.274652] env[65121]: DEBUG nova.network.neutron [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1204.283935] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1204.283935] env[65121]: value = "task-5107498" [ 1204.283935] env[65121]: _type = "Task" [ 1204.283935] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.301907] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.389590] env[65121]: DEBUG nova.scheduler.client.report [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1204.442931] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107496, 'name': CreateVM_Task, 'duration_secs': 0.491216} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.443142] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1204.443686] env[65121]: WARNING neutronclient.v2_0.client [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1204.444069] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.444227] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1204.444553] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1204.444824] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c1060cd-cb66-45ce-bf77-c4266ef67867 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.450677] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1204.450677] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522b3d47-689b-5eac-6485-65605210ebca" [ 1204.450677] env[65121]: _type = "Task" [ 1204.450677] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.462508] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522b3d47-689b-5eac-6485-65605210ebca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.551654] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107497, 'name': ReconfigVM_Task, 'duration_secs': 0.355285} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.551935] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfigured VM instance instance-0000006c to attach disk [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1204.552221] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance 'f58bf796-69e7-4a61-9cea-78dff8d4fab5' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1204.669881] env[65121]: INFO nova.compute.manager [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Took 17.35 seconds to build instance. [ 1204.782040] env[65121]: WARNING neutronclient.v2_0.client [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1204.782040] env[65121]: WARNING openstack [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.782040] env[65121]: WARNING openstack [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.799818] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107498, 'name': PowerOffVM_Task, 'duration_secs': 0.334938} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.800391] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1204.801410] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb5a995-3f6c-440d-8a7e-5086a56af723 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.827054] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeda94fb-675c-4f07-aac9-b6f5783c943d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.895420] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.963049] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522b3d47-689b-5eac-6485-65605210ebca, 'name': SearchDatastore_Task, 'duration_secs': 0.012381} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.963609] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1204.963780] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1204.963966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.964140] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1204.964323] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1204.964597] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-baa6235c-3b28-4d28-999e-8be944aefb86 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.978084] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1204.978084] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1204.978573] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffbaadfa-d4b9-4ca9-8c00-4f40c2026638 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.985528] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1204.985528] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525bee61-8781-d477-f4ae-ff2105afa351" [ 1204.985528] env[65121]: _type = "Task" [ 1204.985528] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.996918] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525bee61-8781-d477-f4ae-ff2105afa351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.061839] env[65121]: DEBUG nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1205.067211] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11afa2a1-0f48-44b1-a8b6-d91ebddda2fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.079866] env[65121]: WARNING openstack [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.079866] env[65121]: WARNING openstack [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.107183] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c26635-e5de-4d8f-bced-2852854f19c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.127931] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance 'f58bf796-69e7-4a61-9cea-78dff8d4fab5' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1205.138118] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1205.138383] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1205.138516] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1205.138810] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1205.138898] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1205.138958] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1205.139209] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1205.139369] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1205.139528] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1205.139686] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1205.139851] env[65121]: DEBUG nova.virt.hardware [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1205.140781] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b003f6b3-2a94-456e-b904-976b006e8bde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.150352] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fb6337-e537-42c7-9cd2-a56a68e4f284 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.172401] env[65121]: DEBUG oslo_concurrency.lockutils [None req-577b18e6-57ba-4bc1-a37f-87bc078a40c0 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.867s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.228654] env[65121]: WARNING neutronclient.v2_0.client [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.229411] env[65121]: WARNING openstack [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.229752] env[65121]: WARNING openstack [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.339656] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1205.339992] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ad892e76-d98a-4dc3-83ea-192cbed37349 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.348129] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1205.348129] env[65121]: value = "task-5107500" [ 1205.348129] env[65121]: _type = "Task" [ 1205.348129] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.350410] env[65121]: DEBUG nova.network.neutron [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updated VIF entry in instance network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1205.350718] env[65121]: DEBUG nova.network.neutron [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": "3cbde80d-7df5-4785-821e-b45797716ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1205.362105] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107500, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.420309] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6044ca1a-2c05-4c0a-ade7-f39f1b698eb1 tempest-ServerActionsV293TestJSON-683979676 tempest-ServerActionsV293TestJSON-683979676-project-member] Lock "4f86672c-425d-4641-a68c-2d6f5324e382" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.851s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.499799] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525bee61-8781-d477-f4ae-ff2105afa351, 'name': SearchDatastore_Task, 'duration_secs': 0.011358} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.500360] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ebe0382-c592-4f9e-b0c3-80d47bbaf217 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.507708] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1205.507708] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c9abc9-a914-9d78-c2f7-86b031f26fd4" [ 1205.507708] env[65121]: _type = "Task" [ 1205.507708] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.520428] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c9abc9-a914-9d78-c2f7-86b031f26fd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.646596] env[65121]: WARNING neutronclient.v2_0.client [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.694303] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.694549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.697029] env[65121]: DEBUG nova.network.neutron [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Port 99be1255-3e71-41e6-936a-acbf44bf9de9 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1205.767166] env[65121]: DEBUG nova.network.neutron [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Successfully updated port: c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1205.858127] env[65121]: DEBUG oslo_concurrency.lockutils [req-977c5b89-7d76-46b0-8714-0a9a1dfaf801 req-06e83ea3-6ff4-4fc3-9960-3b1b0fa9de35 service nova] Releasing lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1205.866140] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107500, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.973053] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.973403] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.973484] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.973678] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.973839] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.976354] env[65121]: INFO nova.compute.manager [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Terminating instance [ 1206.019994] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c9abc9-a914-9d78-c2f7-86b031f26fd4, 'name': SearchDatastore_Task, 'duration_secs': 0.016832} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.020368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1206.020640] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1206.020906] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c40934a3-1c49-44a4-8de1-ba1b39606838 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.029529] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1206.029529] env[65121]: value = "task-5107501" [ 1206.029529] env[65121]: _type = "Task" [ 1206.029529] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.039104] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107501, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.205986] env[65121]: DEBUG nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1206.272621] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.272621] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquired lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1206.272621] env[65121]: DEBUG nova.network.neutron [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1206.308211] env[65121]: DEBUG nova.compute.manager [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Received event network-vif-plugged-c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1206.308500] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Acquiring lock "e19d6d9b-d15d-4780-8564-f479573ce59b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.308713] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.308924] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.309290] env[65121]: DEBUG nova.compute.manager [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] No waiting events found dispatching network-vif-plugged-c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1206.309378] env[65121]: WARNING nova.compute.manager [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Received unexpected event network-vif-plugged-c6e3761c-d9e9-4e36-937f-2df79e305a63 for instance with vm_state building and task_state spawning. [ 1206.309609] env[65121]: DEBUG nova.compute.manager [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Received event network-changed-c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1206.309781] env[65121]: DEBUG nova.compute.manager [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Refreshing instance network info cache due to event network-changed-c6e3761c-d9e9-4e36-937f-2df79e305a63. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1206.309973] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Acquiring lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.368282] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107500, 'name': CreateSnapshot_Task, 'duration_secs': 0.870625} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.368647] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1206.369967] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beafc6fb-29fe-4eaa-ba4b-e9063ab5545a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.480853] env[65121]: DEBUG nova.compute.manager [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1206.481170] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1206.482388] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0486039-0468-48e7-9bc7-84b3b0cf3cc9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.492484] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1206.492877] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d0d9db7-bd12-431d-a4fe-fe361c4efb0d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.501618] env[65121]: DEBUG oslo_vmware.api [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1206.501618] env[65121]: value = "task-5107502" [ 1206.501618] env[65121]: _type = "Task" [ 1206.501618] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.513207] env[65121]: DEBUG oslo_vmware.api [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107502, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.539628] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107501, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486962} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.539989] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1206.540275] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1206.540621] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3173cbd7-7a29-4c3f-98fe-d2e33f2b7335 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.551822] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1206.551822] env[65121]: value = "task-5107503" [ 1206.551822] env[65121]: _type = "Task" [ 1206.551822] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.562976] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.737232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.737232] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.737405] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.759842] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.760175] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1206.762500] env[65121]: INFO nova.compute.claims [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1206.775433] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.776891] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.820681] env[65121]: DEBUG nova.network.neutron [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1206.842693] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.843155] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.894516] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1206.895557] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ab4238a9-237c-4190-bf28-6da7283ebd1a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.904913] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1206.904913] env[65121]: value = "task-5107505" [ 1206.904913] env[65121]: _type = "Task" [ 1206.904913] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.914890] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107505, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.927606] env[65121]: WARNING neutronclient.v2_0.client [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.928748] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.929198] env[65121]: WARNING openstack [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.012448] env[65121]: DEBUG oslo_vmware.api [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107502, 'name': PowerOffVM_Task, 'duration_secs': 0.317017} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.012853] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1207.012853] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1207.013074] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3cda8ea1-9ca6-4cd7-8d27-7219ec5aaa5a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.031507] env[65121]: DEBUG nova.network.neutron [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Updating instance_info_cache with network_info: [{"id": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "address": "fa:16:3e:04:d2:b6", "network": {"id": "98611cce-821d-4222-96a4-187a62b96382", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-28779978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b07f57e89c774ef1bbaff256800ab80b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6e3761c-d9", "ovs_interfaceid": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1207.063167] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078998} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.063440] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1207.064365] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fb45e7-bafd-48c2-b87c-34a0fb13aa8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.093016] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1207.095259] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8355ee09-525a-49ab-9d3b-dfacb74c68d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.112302] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1207.112557] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1207.112715] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Deleting the datastore file [datastore1] fb69eaef-32e9-4ade-9ed2-27ed1df36ce3 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1207.113040] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb0bb1dc-def6-4215-8d88-fe88bd3b0aba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.122829] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1207.122829] env[65121]: value = "task-5107508" [ 1207.122829] env[65121]: _type = "Task" [ 1207.122829] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.124924] env[65121]: DEBUG oslo_vmware.api [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for the task: (returnval){ [ 1207.124924] env[65121]: value = "task-5107507" [ 1207.124924] env[65121]: _type = "Task" [ 1207.124924] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.139875] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107508, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.143631] env[65121]: DEBUG oslo_vmware.api [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107507, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.417785] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107505, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.536352] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Releasing lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1207.536352] env[65121]: DEBUG nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Instance network_info: |[{"id": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "address": "fa:16:3e:04:d2:b6", "network": {"id": "98611cce-821d-4222-96a4-187a62b96382", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-28779978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b07f57e89c774ef1bbaff256800ab80b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6e3761c-d9", "ovs_interfaceid": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1207.536352] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Acquired lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1207.536352] env[65121]: DEBUG nova.network.neutron [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Refreshing network info cache for port c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1207.537550] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:d2:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1931669-8959-4e86-a603-e206bcf2b47a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6e3761c-d9e9-4e36-937f-2df79e305a63', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1207.550023] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Creating folder: Project (b07f57e89c774ef1bbaff256800ab80b). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1207.552118] env[65121]: WARNING neutronclient.v2_0.client [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.552903] env[65121]: WARNING openstack [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.553339] env[65121]: WARNING openstack [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.565162] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01557681-429d-4949-9af3-c7cd45020add {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.580302] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Created folder: Project (b07f57e89c774ef1bbaff256800ab80b) in parent group-v993268. [ 1207.580537] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Creating folder: Instances. Parent ref: group-v993590. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1207.580808] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d2c25ab-de89-4759-9ec9-f573e483c665 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.599577] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Created folder: Instances in parent group-v993590. [ 1207.599857] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1207.602510] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1207.603512] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f54bae7d-695d-46bf-b81f-55d863f0e381 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.625103] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1207.625103] env[65121]: value = "task-5107511" [ 1207.625103] env[65121]: _type = "Task" [ 1207.625103] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.643068] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107508, 'name': ReconfigVM_Task, 'duration_secs': 0.444578} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.650081] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Reconfigured VM instance instance-00000078 to attach disk [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1207.650271] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107511, 'name': CreateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.650437] env[65121]: DEBUG oslo_vmware.api [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Task: {'id': task-5107507, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.297876} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.650640] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3520943-37c2-4ef9-8ec5-36357df96334 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.652462] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1207.653945] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1207.653945] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1207.653945] env[65121]: INFO nova.compute.manager [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1207.653945] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1207.653945] env[65121]: DEBUG nova.compute.manager [-] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1207.653945] env[65121]: DEBUG nova.network.neutron [-] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1207.653945] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.655723] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.655723] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.672678] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1207.672678] env[65121]: value = "task-5107512" [ 1207.672678] env[65121]: _type = "Task" [ 1207.672678] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.682749] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107512, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.717464] env[65121]: WARNING openstack [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.717884] env[65121]: WARNING openstack [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.727386] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.749959] env[65121]: WARNING neutronclient.v2_0.client [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.924025] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107505, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.038126] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88414b33-99c7-4db8-83df-e98b7c5caa06 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.048442] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.048644] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1208.048844] env[65121]: DEBUG nova.network.neutron [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1208.053013] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299cf96c-88fd-450d-9695-a8aa084eb008 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.096407] env[65121]: WARNING neutronclient.v2_0.client [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.100356] env[65121]: WARNING openstack [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.100860] env[65121]: WARNING openstack [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.110553] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbfc457-3a41-4e6a-8664-7c23a2a7df0a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.120584] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa248fe-f754-4ab8-8be3-18e66b41aed9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.136184] env[65121]: DEBUG nova.compute.provider_tree [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1208.147867] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107511, 'name': CreateVM_Task, 'duration_secs': 0.372848} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.148547] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1208.153572] env[65121]: WARNING neutronclient.v2_0.client [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.153912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.154080] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1208.154389] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1208.156304] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12f7c06d-ed30-4ccc-b0e2-3347af29cc88 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.162470] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1208.162470] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5216148a-6715-4762-2f56-714d56ee17a6" [ 1208.162470] env[65121]: _type = "Task" [ 1208.162470] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.173388] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5216148a-6715-4762-2f56-714d56ee17a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.182249] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107512, 'name': Rename_Task, 'duration_secs': 0.181351} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.182538] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1208.182782] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4a8db2b-8128-446f-a371-f5c551eafcbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.190248] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1208.190248] env[65121]: value = "task-5107513" [ 1208.190248] env[65121]: _type = "Task" [ 1208.190248] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.197048] env[65121]: DEBUG nova.network.neutron [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Updated VIF entry in instance network info cache for port c6e3761c-d9e9-4e36-937f-2df79e305a63. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1208.197414] env[65121]: DEBUG nova.network.neutron [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Updating instance_info_cache with network_info: [{"id": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "address": "fa:16:3e:04:d2:b6", "network": {"id": "98611cce-821d-4222-96a4-187a62b96382", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-28779978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b07f57e89c774ef1bbaff256800ab80b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6e3761c-d9", "ovs_interfaceid": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1208.202460] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.338807] env[65121]: DEBUG nova.compute.manager [req-f526cfb6-7f7b-4559-a4c9-39745657b53b req-bab1b225-a742-44a8-b65a-8ba00acaf61b service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Received event network-vif-deleted-c59158a9-d9f6-4465-b99a-724753944d6a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1208.338894] env[65121]: INFO nova.compute.manager [req-f526cfb6-7f7b-4559-a4c9-39745657b53b req-bab1b225-a742-44a8-b65a-8ba00acaf61b service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Neutron deleted interface c59158a9-d9f6-4465-b99a-724753944d6a; detaching it from the instance and deleting it from the info cache [ 1208.339118] env[65121]: DEBUG nova.network.neutron [req-f526cfb6-7f7b-4559-a4c9-39745657b53b req-bab1b225-a742-44a8-b65a-8ba00acaf61b service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1208.419753] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107505, 'name': CloneVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.523511] env[65121]: DEBUG nova.network.neutron [-] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1208.560456] env[65121]: WARNING neutronclient.v2_0.client [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.561081] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.561735] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.642925] env[65121]: DEBUG nova.scheduler.client.report [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1208.682916] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5216148a-6715-4762-2f56-714d56ee17a6, 'name': SearchDatastore_Task, 'duration_secs': 0.010578} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.682916] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1208.683037] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1208.683240] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.683396] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1208.684048] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1208.684048] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5abf05c-250d-47f9-8137-d96203b400e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.703418] env[65121]: DEBUG oslo_concurrency.lockutils [req-9a25fcf7-2426-420c-914a-0d0662fe4732 req-a734fa87-e3df-44a1-9811-a81e95badb74 service nova] Releasing lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1208.703875] env[65121]: DEBUG oslo_vmware.api [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107513, 'name': PowerOnVM_Task, 'duration_secs': 0.511444} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.704251] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1208.704656] env[65121]: INFO nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Took 7.91 seconds to spawn the instance on the hypervisor. [ 1208.704813] env[65121]: DEBUG nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1208.707082] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e4c453-4be6-4d4d-8066-6cac3015ef87 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.709991] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1208.710295] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1208.711448] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78e5c1c3-d3c6-40bf-b6b0-760b31b32293 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.720720] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1208.720720] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5265ec96-3625-2d9a-6a94-922cc723cccc" [ 1208.720720] env[65121]: _type = "Task" [ 1208.720720] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.733424] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5265ec96-3625-2d9a-6a94-922cc723cccc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.753889] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.754274] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.844130] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc5a5ba0-e1a5-480e-be04-4acfe0548889 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.847942] env[65121]: WARNING neutronclient.v2_0.client [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.849493] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.849493] env[65121]: WARNING openstack [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.869879] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7ecd87-5937-49ff-9d8b-38d0925b941e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.927890] env[65121]: DEBUG nova.compute.manager [req-f526cfb6-7f7b-4559-a4c9-39745657b53b req-bab1b225-a742-44a8-b65a-8ba00acaf61b service nova] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Detach interface failed, port_id=c59158a9-d9f6-4465-b99a-724753944d6a, reason: Instance fb69eaef-32e9-4ade-9ed2-27ed1df36ce3 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1208.949039] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107505, 'name': CloneVM_Task, 'duration_secs': 1.568861} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.949379] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Created linked-clone VM from snapshot [ 1208.950616] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8743f56-d8d2-4687-b7ba-fb320515e56d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.962991] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Uploading image ceda632d-08b0-45f0-ae23-8456ddb4144e {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1208.980444] env[65121]: DEBUG nova.network.neutron [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1208.994849] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1208.994849] env[65121]: value = "vm-993589" [ 1208.994849] env[65121]: _type = "VirtualMachine" [ 1208.994849] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1208.995573] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-99771b42-0702-4d80-8270-f74689e8b3f8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.003275] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease: (returnval){ [ 1209.003275] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fcaa99-ad68-9cc1-1021-83f6220a419a" [ 1209.003275] env[65121]: _type = "HttpNfcLease" [ 1209.003275] env[65121]: } obtained for exporting VM: (result){ [ 1209.003275] env[65121]: value = "vm-993589" [ 1209.003275] env[65121]: _type = "VirtualMachine" [ 1209.003275] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1209.003568] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the lease: (returnval){ [ 1209.003568] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fcaa99-ad68-9cc1-1021-83f6220a419a" [ 1209.003568] env[65121]: _type = "HttpNfcLease" [ 1209.003568] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1209.012116] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1209.012116] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fcaa99-ad68-9cc1-1021-83f6220a419a" [ 1209.012116] env[65121]: _type = "HttpNfcLease" [ 1209.012116] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1209.027172] env[65121]: INFO nova.compute.manager [-] [instance: fb69eaef-32e9-4ade-9ed2-27ed1df36ce3] Took 1.37 seconds to deallocate network for instance. [ 1209.149280] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.149865] env[65121]: DEBUG nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1209.238642] env[65121]: INFO nova.compute.manager [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Took 12.37 seconds to build instance. [ 1209.243539] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5265ec96-3625-2d9a-6a94-922cc723cccc, 'name': SearchDatastore_Task, 'duration_secs': 0.025015} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.244872] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7162f426-30ec-40cd-afba-9f6a851addb2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.251825] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1209.251825] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52313954-54ca-411e-74f4-f86bcc7abe04" [ 1209.251825] env[65121]: _type = "Task" [ 1209.251825] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.263319] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52313954-54ca-411e-74f4-f86bcc7abe04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.485098] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1209.513674] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1209.513674] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fcaa99-ad68-9cc1-1021-83f6220a419a" [ 1209.513674] env[65121]: _type = "HttpNfcLease" [ 1209.513674] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1209.514200] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1209.514200] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52fcaa99-ad68-9cc1-1021-83f6220a419a" [ 1209.514200] env[65121]: _type = "HttpNfcLease" [ 1209.514200] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1209.514827] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f0661c-6f3a-4b8c-beca-36034a25c062 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.525617] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ef6fb-e1ff-b265-84b0-c2b1f31b2699/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1209.525617] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ef6fb-e1ff-b265-84b0-c2b1f31b2699/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1209.589234] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.589529] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.589751] env[65121]: DEBUG nova.objects.instance [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lazy-loading 'resources' on Instance uuid fb69eaef-32e9-4ade-9ed2-27ed1df36ce3 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1209.631390] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b5419432-5c71-4c70-8ef1-92d79217e285 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.655398] env[65121]: DEBUG nova.compute.utils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1209.656897] env[65121]: DEBUG nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1209.657725] env[65121]: DEBUG nova.network.neutron [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1209.657725] env[65121]: WARNING neutronclient.v2_0.client [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.658267] env[65121]: WARNING neutronclient.v2_0.client [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.659163] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.659535] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.740774] env[65121]: DEBUG nova.policy [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e20599b4b4f4459eba25b658da1a8484', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a23c876483f34b1db8beec9a669f7822', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1209.744445] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8a44b24d-f10a-417b-864c-6fd61ad01941 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.885s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.763696] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52313954-54ca-411e-74f4-f86bcc7abe04, 'name': SearchDatastore_Task, 'duration_secs': 0.014036} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.763962] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1209.764239] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e19d6d9b-d15d-4780-8564-f479573ce59b/e19d6d9b-d15d-4780-8564-f479573ce59b.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1209.764524] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d7f2d37-e06c-4d8e-a76d-59c2bb46d4ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.773806] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1209.773806] env[65121]: value = "task-5107516" [ 1209.773806] env[65121]: _type = "Task" [ 1209.773806] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.784929] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107516, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.025020] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8c673b-c489-4388-80bf-01520f3ed3e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.053095] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad26390b-2b56-4dfd-9d23-8b00b760c851 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.065936] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance 'f58bf796-69e7-4a61-9cea-78dff8d4fab5' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1210.174689] env[65121]: DEBUG nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1210.204397] env[65121]: DEBUG nova.network.neutron [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Successfully created port: 1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1210.295194] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107516, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.398375] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867e2ef6-fd5d-4afb-ae53-b1e5d265b1d0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.407676] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d88256-4b27-4f55-bbe0-8c1ca7afe311 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.454972] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bfacaf-15b3-40dc-85b2-4f3fb5103e52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.464472] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5f8204-043c-4d2a-9648-053a9faeae7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.482783] env[65121]: DEBUG nova.compute.provider_tree [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1210.575853] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1210.576560] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fffe6d7-c031-4e75-98ae-9444acc40f8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.585604] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1210.585604] env[65121]: value = "task-5107517" [ 1210.585604] env[65121]: _type = "Task" [ 1210.585604] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.595547] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107517, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.775021] env[65121]: DEBUG nova.compute.manager [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1210.775229] env[65121]: DEBUG nova.compute.manager [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing instance network info cache due to event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1210.775618] env[65121]: DEBUG oslo_concurrency.lockutils [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Acquiring lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.775878] env[65121]: DEBUG oslo_concurrency.lockutils [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Acquired lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1210.776095] env[65121]: DEBUG nova.network.neutron [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1210.789707] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107516, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644391} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.790761] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore2] e19d6d9b-d15d-4780-8564-f479573ce59b/e19d6d9b-d15d-4780-8564-f479573ce59b.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1210.791058] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1210.791421] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2b0d1bb-eac5-495e-9911-0fd6282bed94 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.799980] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1210.799980] env[65121]: value = "task-5107518" [ 1210.799980] env[65121]: _type = "Task" [ 1210.799980] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.815540] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.987047] env[65121]: DEBUG nova.scheduler.client.report [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1211.096907] env[65121]: DEBUG oslo_vmware.api [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107517, 'name': PowerOnVM_Task, 'duration_secs': 0.493609} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.097311] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1211.097539] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-e8a88945-53eb-480b-9c75-b2424f8fae26 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance 'f58bf796-69e7-4a61-9cea-78dff8d4fab5' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1211.194108] env[65121]: DEBUG nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1211.219813] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1211.220671] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1211.220671] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1211.220962] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1211.221259] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1211.221362] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1211.221587] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1211.221930] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1211.222112] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1211.222329] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1211.222628] env[65121]: DEBUG nova.virt.hardware [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1211.223798] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc390c74-7fd7-4db6-a071-3047afe9d983 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.233334] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fb0d7c-28de-4518-8ef7-146232beb8f6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.279964] env[65121]: WARNING neutronclient.v2_0.client [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1211.280624] env[65121]: WARNING openstack [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.281255] env[65121]: WARNING openstack [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.311884] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07243} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.312335] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1211.313353] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b3fb28-f6b5-4cfb-ba8b-bbda3de5641b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.339854] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] e19d6d9b-d15d-4780-8564-f479573ce59b/e19d6d9b-d15d-4780-8564-f479573ce59b.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1211.342917] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1622c387-5e07-48a2-ad3b-13af6406ff62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.365669] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1211.365669] env[65121]: value = "task-5107520" [ 1211.365669] env[65121]: _type = "Task" [ 1211.365669] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.375810] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107520, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.456901] env[65121]: WARNING openstack [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.457444] env[65121]: WARNING openstack [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.494421] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.526313] env[65121]: INFO nova.scheduler.client.report [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Deleted allocations for instance fb69eaef-32e9-4ade-9ed2-27ed1df36ce3 [ 1211.550814] env[65121]: WARNING neutronclient.v2_0.client [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1211.550814] env[65121]: WARNING openstack [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.551388] env[65121]: WARNING openstack [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.840020] env[65121]: DEBUG nova.network.neutron [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updated VIF entry in instance network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1211.840020] env[65121]: DEBUG nova.network.neutron [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": "3cbde80d-7df5-4785-821e-b45797716ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1211.881178] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107520, 'name': ReconfigVM_Task, 'duration_secs': 0.493313} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.883464] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Reconfigured VM instance instance-00000079 to attach disk [datastore2] e19d6d9b-d15d-4780-8564-f479573ce59b/e19d6d9b-d15d-4780-8564-f479573ce59b.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1211.885882] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e651773e-45ff-443b-a10b-10cd005481a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.894653] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1211.894653] env[65121]: value = "task-5107521" [ 1211.894653] env[65121]: _type = "Task" [ 1211.894653] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.909454] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107521, 'name': Rename_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.973932] env[65121]: DEBUG nova.network.neutron [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Successfully updated port: 1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1211.989676] env[65121]: DEBUG nova.compute.manager [req-29d44131-1e85-4412-9d54-1b0ff8a291e4 req-a1684422-5f04-4a6b-a2ed-ce317994bea1 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Received event network-vif-plugged-1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1211.989849] env[65121]: DEBUG oslo_concurrency.lockutils [req-29d44131-1e85-4412-9d54-1b0ff8a291e4 req-a1684422-5f04-4a6b-a2ed-ce317994bea1 service nova] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.990280] env[65121]: DEBUG oslo_concurrency.lockutils [req-29d44131-1e85-4412-9d54-1b0ff8a291e4 req-a1684422-5f04-4a6b-a2ed-ce317994bea1 service nova] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.990528] env[65121]: DEBUG oslo_concurrency.lockutils [req-29d44131-1e85-4412-9d54-1b0ff8a291e4 req-a1684422-5f04-4a6b-a2ed-ce317994bea1 service nova] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.990802] env[65121]: DEBUG nova.compute.manager [req-29d44131-1e85-4412-9d54-1b0ff8a291e4 req-a1684422-5f04-4a6b-a2ed-ce317994bea1 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] No waiting events found dispatching network-vif-plugged-1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1211.991098] env[65121]: WARNING nova.compute.manager [req-29d44131-1e85-4412-9d54-1b0ff8a291e4 req-a1684422-5f04-4a6b-a2ed-ce317994bea1 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Received unexpected event network-vif-plugged-1c9a1903-516b-4056-ba71-cf778ceebf76 for instance with vm_state building and task_state spawning. [ 1212.036705] env[65121]: DEBUG oslo_concurrency.lockutils [None req-53f75b1f-d97d-496f-b2ec-bf7a273661f2 tempest-ServerAddressesTestJSON-842682187 tempest-ServerAddressesTestJSON-842682187-project-member] Lock "fb69eaef-32e9-4ade-9ed2-27ed1df36ce3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.063s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1212.341604] env[65121]: DEBUG oslo_concurrency.lockutils [req-c322b358-89fc-444f-8ebe-25ddbcc23e18 req-77187b43-9a94-4709-8403-92e638881891 service nova] Releasing lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1212.408823] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107521, 'name': Rename_Task, 'duration_secs': 0.232066} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.408823] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1212.409149] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d205b195-f384-4802-9a91-12b085f0ce7a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.419048] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1212.419048] env[65121]: value = "task-5107522" [ 1212.419048] env[65121]: _type = "Task" [ 1212.419048] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.435722] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.477078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.477078] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.477078] env[65121]: DEBUG nova.network.neutron [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1212.931184] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107522, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.981196] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.981196] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.060804] env[65121]: DEBUG nova.network.neutron [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1213.105269] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.105801] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.216477] env[65121]: WARNING neutronclient.v2_0.client [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.217617] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.218477] env[65121]: WARNING openstack [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.384802] env[65121]: DEBUG nova.network.neutron [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.432500] env[65121]: DEBUG oslo_vmware.api [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107522, 'name': PowerOnVM_Task, 'duration_secs': 0.676536} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.433255] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1213.433255] env[65121]: INFO nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Took 8.37 seconds to spawn the instance on the hypervisor. [ 1213.433405] env[65121]: DEBUG nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1213.434902] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae392b2f-480e-48d2-acfd-29b6391c0801 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.668924] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.708718] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.709158] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.746494] env[65121]: DEBUG nova.network.neutron [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Port 99be1255-3e71-41e6-936a-acbf44bf9de9 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1213.746806] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.746961] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.747403] env[65121]: DEBUG nova.network.neutron [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1213.888809] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.889266] env[65121]: DEBUG nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Instance network_info: |[{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1213.890208] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:99:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '12d8eedb-97cb-4d3b-b364-42d7fd8b3c85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c9a1903-516b-4056-ba71-cf778ceebf76', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1213.901353] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1213.901750] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1213.902167] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a53f5626-b147-4548-85d1-f3668d257774 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.926767] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1213.926767] env[65121]: value = "task-5107524" [ 1213.926767] env[65121]: _type = "Task" [ 1213.926767] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.937008] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107524, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.956056] env[65121]: INFO nova.compute.manager [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Took 14.77 seconds to build instance. [ 1214.022531] env[65121]: DEBUG nova.compute.manager [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Received event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1214.022783] env[65121]: DEBUG nova.compute.manager [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing instance network info cache due to event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1214.023018] env[65121]: DEBUG oslo_concurrency.lockutils [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Acquiring lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.023264] env[65121]: DEBUG oslo_concurrency.lockutils [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Acquired lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1214.023458] env[65121]: DEBUG nova.network.neutron [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1214.252334] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.252334] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.252334] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.439626] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107524, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.459029] env[65121]: DEBUG oslo_concurrency.lockutils [None req-15fcfa8e-5d40-4aac-9dac-b534b667680d tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.296s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1214.472869] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.473421] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.526964] env[65121]: WARNING neutronclient.v2_0.client [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.528141] env[65121]: WARNING openstack [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.528607] env[65121]: WARNING openstack [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.686550] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.687343] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.687714] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.772610] env[65121]: WARNING openstack [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.773052] env[65121]: WARNING openstack [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.789049] env[65121]: DEBUG nova.compute.manager [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Received event network-changed-c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1214.789049] env[65121]: DEBUG nova.compute.manager [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Refreshing instance network info cache due to event network-changed-c6e3761c-d9e9-4e36-937f-2df79e305a63. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1214.789289] env[65121]: DEBUG oslo_concurrency.lockutils [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Acquiring lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.789403] env[65121]: DEBUG oslo_concurrency.lockutils [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Acquired lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1214.789558] env[65121]: DEBUG nova.network.neutron [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Refreshing network info cache for port c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1214.824529] env[65121]: DEBUG nova.network.neutron [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1214.851018] env[65121]: WARNING neutronclient.v2_0.client [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.851845] env[65121]: WARNING openstack [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.852231] env[65121]: WARNING openstack [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.944111] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107524, 'name': CreateVM_Task, 'duration_secs': 0.59818} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.947187] env[65121]: DEBUG nova.network.neutron [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updated VIF entry in instance network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1214.947701] env[65121]: DEBUG nova.network.neutron [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1214.949221] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1214.949790] env[65121]: WARNING neutronclient.v2_0.client [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.950195] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.950420] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1214.950772] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1214.951326] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc1d6383-311f-4079-9b8e-b23b50759e7e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.957587] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1214.957587] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52d2f21f-9d33-69c4-0422-880758489fb0" [ 1214.957587] env[65121]: _type = "Task" [ 1214.957587] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.969759] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d2f21f-9d33-69c4-0422-880758489fb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.292887] env[65121]: WARNING neutronclient.v2_0.client [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.293982] env[65121]: WARNING openstack [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1215.294614] env[65121]: WARNING openstack [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1215.327966] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.421508] env[65121]: WARNING openstack [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1215.421994] env[65121]: WARNING openstack [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1215.452911] env[65121]: DEBUG oslo_concurrency.lockutils [req-b614501c-2aeb-430b-89ee-e95268e5a074 req-2c22b391-e87b-45fc-be7d-3cfa2afea550 service nova] Releasing lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.469882] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52d2f21f-9d33-69c4-0422-880758489fb0, 'name': SearchDatastore_Task, 'duration_secs': 0.022519} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.470275] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.470536] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1215.470790] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.470928] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1215.471115] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1215.471389] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97e6d5b8-5b47-43ab-ac8f-ccf48b3f4ff8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.482445] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1215.482668] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1215.483415] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-638b4232-ca75-4cc1-89cd-9f5602011b3f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.489915] env[65121]: WARNING neutronclient.v2_0.client [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.490803] env[65121]: WARNING openstack [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1215.491231] env[65121]: WARNING openstack [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1215.505721] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1215.505721] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5291358a-d4c7-cad6-64a0-a17add171702" [ 1215.505721] env[65121]: _type = "Task" [ 1215.505721] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.514884] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5291358a-d4c7-cad6-64a0-a17add171702, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.583533] env[65121]: DEBUG nova.network.neutron [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Updated VIF entry in instance network info cache for port c6e3761c-d9e9-4e36-937f-2df79e305a63. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1215.583968] env[65121]: DEBUG nova.network.neutron [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Updating instance_info_cache with network_info: [{"id": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "address": "fa:16:3e:04:d2:b6", "network": {"id": "98611cce-821d-4222-96a4-187a62b96382", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-28779978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b07f57e89c774ef1bbaff256800ab80b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6e3761c-d9", "ovs_interfaceid": "c6e3761c-d9e9-4e36-937f-2df79e305a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1215.831516] env[65121]: DEBUG nova.compute.manager [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65121) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1215.831733] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.831968] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1216.016920] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5291358a-d4c7-cad6-64a0-a17add171702, 'name': SearchDatastore_Task, 'duration_secs': 0.020178} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.017817] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1df8b83-7593-4300-a788-d6909e39c670 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.024218] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1216.024218] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]520a3b14-4b1f-227b-cfd7-93293c923804" [ 1216.024218] env[65121]: _type = "Task" [ 1216.024218] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.033645] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520a3b14-4b1f-227b-cfd7-93293c923804, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.086745] env[65121]: DEBUG oslo_concurrency.lockutils [req-96062ca7-7a26-4095-b507-e1cefb1b7950 req-396cba76-920c-44fe-b48c-0fe32c61dfc4 service nova] Releasing lock "refresh_cache-e19d6d9b-d15d-4780-8564-f479573ce59b" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1216.334786] env[65121]: DEBUG nova.objects.instance [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'migration_context' on Instance uuid f58bf796-69e7-4a61-9cea-78dff8d4fab5 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.536484] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]520a3b14-4b1f-227b-cfd7-93293c923804, 'name': SearchDatastore_Task, 'duration_secs': 0.014906} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.536819] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1216.537083] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/d34b8457-adc1-4d6b-befd-f6f1aff6d1fb.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1216.537362] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-501fca1a-580f-4f7b-bc24-9f99e5ac10fd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.544838] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1216.544838] env[65121]: value = "task-5107526" [ 1216.544838] env[65121]: _type = "Task" [ 1216.544838] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.553938] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.040459] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4fce46-958c-4bc6-a669-8cf5c1cd49ac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.053461] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd45b0d6-c323-432a-a5f4-15faef5d823b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.060241] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107526, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.091536] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1274d23c-8d48-447c-be4f-d005906cc3b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.100635] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f6b396-60b5-4027-9305-09e7b47a8da2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.121876] env[65121]: DEBUG nova.compute.provider_tree [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.556835] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660697} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.557155] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/d34b8457-adc1-4d6b-befd-f6f1aff6d1fb.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1217.557429] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1217.557680] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0aaf5084-cbd8-40f3-91d2-a166687e1cba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.564998] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1217.564998] env[65121]: value = "task-5107527" [ 1217.564998] env[65121]: _type = "Task" [ 1217.564998] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.573735] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.626116] env[65121]: DEBUG nova.scheduler.client.report [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1218.076359] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094227} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.076647] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1218.077538] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf8b113c-4b24-428f-8ef3-13386433812b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.102670] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/d34b8457-adc1-4d6b-befd-f6f1aff6d1fb.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1218.103024] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41f96766-feb7-4dde-b8f9-11d808105876 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.124561] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1218.124561] env[65121]: value = "task-5107528" [ 1218.124561] env[65121]: _type = "Task" [ 1218.124561] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.138919] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107528, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.332672] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ef6fb-e1ff-b265-84b0-c2b1f31b2699/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1218.333670] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6f9fb5-ada5-4135-8c66-d793d678a29f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.340506] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ef6fb-e1ff-b265-84b0-c2b1f31b2699/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1218.340680] env[65121]: ERROR oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ef6fb-e1ff-b265-84b0-c2b1f31b2699/disk-0.vmdk due to incomplete transfer. [ 1218.340913] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e356c56d-0e51-4190-896c-c0813a7699da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.350789] env[65121]: DEBUG oslo_vmware.rw_handles [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528ef6fb-e1ff-b265-84b0-c2b1f31b2699/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1218.350994] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Uploaded image ceda632d-08b0-45f0-ae23-8456ddb4144e to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1218.353808] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1218.354115] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-31693780-ab08-4a57-9820-b2dc94874028 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.361336] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1218.361336] env[65121]: value = "task-5107529" [ 1218.361336] env[65121]: _type = "Task" [ 1218.361336] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.370927] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107529, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.638726] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.641594] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.809s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.871447] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107529, 'name': Destroy_Task, 'duration_secs': 0.358034} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.871775] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Destroyed the VM [ 1218.872019] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1218.872312] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-76391e35-5e83-4f08-bcc7-04ffa4702958 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.879661] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1218.879661] env[65121]: value = "task-5107530" [ 1218.879661] env[65121]: _type = "Task" [ 1218.879661] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.889915] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107530, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.078372] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "91a36110-5e97-4553-90c3-3e4213e30c14" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1219.078731] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.136271] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107528, 'name': ReconfigVM_Task, 'duration_secs': 0.853433} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.136556] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfigured VM instance instance-0000007a to attach disk [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/d34b8457-adc1-4d6b-befd-f6f1aff6d1fb.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1219.137244] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cec90c81-748c-42c0-82b4-9a8d623f6571 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.144488] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1219.144488] env[65121]: value = "task-5107531" [ 1219.144488] env[65121]: _type = "Task" [ 1219.144488] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.153616] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107531, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.390372] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107530, 'name': RemoveSnapshot_Task, 'duration_secs': 0.377863} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.390634] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1219.390910] env[65121]: DEBUG nova.compute.manager [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1219.391715] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb3b429-3d4a-4c95-be13-795c3d85d43e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.583207] env[65121]: DEBUG nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1219.655490] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107531, 'name': Rename_Task, 'duration_secs': 0.336274} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.655838] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1219.657967] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19915cac-d6dc-4d92-910b-2b7b0ee518fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.668138] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1219.668138] env[65121]: value = "task-5107532" [ 1219.668138] env[65121]: _type = "Task" [ 1219.668138] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.676736] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.904608] env[65121]: INFO nova.compute.manager [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Shelve offloading [ 1220.107273] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.107516] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.109283] env[65121]: INFO nova.compute.claims [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1220.181612] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107532, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.182766] env[65121]: INFO nova.compute.manager [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Swapping old allocation on dict_keys(['d56783bf-3ede-475a-8c5a-8d8303049e47']) held by migration b70e77be-ee4c-45d8-9e0a-65a78916fa41 for instance [ 1220.208448] env[65121]: DEBUG nova.scheduler.client.report [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Overwriting current allocation {'allocations': {'d56783bf-3ede-475a-8c5a-8d8303049e47': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 172}}, 'project_id': '72237a0d762645588c41231b0a34a796', 'user_id': 'af00bd582c0843949491bbcecbfcd2dc', 'consumer_generation': 1} on consumer f58bf796-69e7-4a61-9cea-78dff8d4fab5 {{(pid=65121) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1220.261441] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1220.294865] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.295176] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1220.295415] env[65121]: DEBUG nova.network.neutron [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1220.408463] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1220.408463] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c905e636-fed2-4ad0-a1c2-4b2de4715690 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.417097] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1220.417097] env[65121]: value = "task-5107533" [ 1220.417097] env[65121]: _type = "Task" [ 1220.417097] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.427907] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1220.428269] env[65121]: DEBUG nova.compute.manager [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1220.429177] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e1ed28-049c-4a57-bb9a-2995e8c8505f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.436749] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.436932] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1220.437142] env[65121]: DEBUG nova.network.neutron [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1220.681121] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107532, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.798573] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1220.799417] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.799817] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1220.925560] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.925942] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1220.940139] env[65121]: WARNING neutronclient.v2_0.client [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1220.940793] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.941152] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.008223] env[65121]: WARNING neutronclient.v2_0.client [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.008901] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.009367] env[65121]: WARNING openstack [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.070779] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.071182] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.113784] env[65121]: DEBUG nova.network.neutron [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [{"id": "99be1255-3e71-41e6-936a-acbf44bf9de9", "address": "fa:16:3e:53:a0:48", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99be1255-3e", "ovs_interfaceid": "99be1255-3e71-41e6-936a-acbf44bf9de9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1221.131856] env[65121]: WARNING neutronclient.v2_0.client [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.132647] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.132989] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.183650] env[65121]: DEBUG oslo_vmware.api [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107532, 'name': PowerOnVM_Task, 'duration_secs': 1.032481} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.183941] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1221.184164] env[65121]: INFO nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Took 9.99 seconds to spawn the instance on the hypervisor. [ 1221.184341] env[65121]: DEBUG nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1221.185161] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67da7d1-797c-4934-980c-64ff5b4fc3d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.233485] env[65121]: DEBUG nova.network.neutron [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91969a04-4a", "ovs_interfaceid": "91969a04-4a8e-4334-829f-3e3ae6828511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1221.313933] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16b671b-99c7-42c9-b7e3-c4bc6fc8f7b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.322124] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689893f9-9150-4e8a-9ecf-269f802ecf64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.356755] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611163c0-86e7-49eb-bee7-129c8376da71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.365826] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e592050-834d-46c6-9d77-1121ece9939c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.380972] env[65121]: DEBUG nova.compute.provider_tree [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.617270] env[65121]: DEBUG oslo_concurrency.lockutils [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-f58bf796-69e7-4a61-9cea-78dff8d4fab5" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1221.617837] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1221.618173] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fb6fa5f-6d8e-466d-addb-deea052ace73 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.626487] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1221.626487] env[65121]: value = "task-5107534" [ 1221.626487] env[65121]: _type = "Task" [ 1221.626487] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.635070] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.712169] env[65121]: INFO nova.compute.manager [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Took 14.98 seconds to build instance. [ 1221.738670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1221.739263] env[65121]: WARNING neutronclient.v2_0.client [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.740136] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.740604] env[65121]: WARNING openstack [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.746491] env[65121]: WARNING neutronclient.v2_0.client [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.884675] env[65121]: DEBUG nova.scheduler.client.report [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1222.046793] env[65121]: DEBUG nova.compute.manager [req-1813e59d-f3e4-46de-b405-4b97faa62622 req-8025f9dc-3f02-4f94-98a5-93aae43235e0 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-vif-unplugged-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1222.047043] env[65121]: DEBUG oslo_concurrency.lockutils [req-1813e59d-f3e4-46de-b405-4b97faa62622 req-8025f9dc-3f02-4f94-98a5-93aae43235e0 service nova] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1222.047295] env[65121]: DEBUG oslo_concurrency.lockutils [req-1813e59d-f3e4-46de-b405-4b97faa62622 req-8025f9dc-3f02-4f94-98a5-93aae43235e0 service nova] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1222.047493] env[65121]: DEBUG oslo_concurrency.lockutils [req-1813e59d-f3e4-46de-b405-4b97faa62622 req-8025f9dc-3f02-4f94-98a5-93aae43235e0 service nova] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1222.047642] env[65121]: DEBUG nova.compute.manager [req-1813e59d-f3e4-46de-b405-4b97faa62622 req-8025f9dc-3f02-4f94-98a5-93aae43235e0 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] No waiting events found dispatching network-vif-unplugged-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1222.047835] env[65121]: WARNING nova.compute.manager [req-1813e59d-f3e4-46de-b405-4b97faa62622 req-8025f9dc-3f02-4f94-98a5-93aae43235e0 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received unexpected event network-vif-unplugged-91969a04-4a8e-4334-829f-3e3ae6828511 for instance with vm_state shelved and task_state shelving_offloading. [ 1222.136877] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107534, 'name': PowerOffVM_Task, 'duration_secs': 0.254079} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.137177] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1222.137960] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1222.138052] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1222.138315] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1222.138356] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1222.138492] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1222.138632] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1222.138828] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1222.139013] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1222.139149] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1222.139358] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1222.139537] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1222.145644] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1222.145930] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d5008f1-7e06-4f6a-90ae-5b7a3088cb68 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.159132] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7754290-5c00-4db6-bf11-aac37634e36e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.168591] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1222.168591] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0890ee19-7ced-47e3-8945-d403b376f1d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.173064] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1222.173064] env[65121]: value = "task-5107535" [ 1222.173064] env[65121]: _type = "Task" [ 1222.173064] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.182941] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107535, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.214738] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c66200c7-8de6-4529-be22-4b882655d885 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.520s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1222.245386] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1222.245601] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1222.245781] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleting the datastore file [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1222.246080] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d1eae61-e48c-49ce-a1b2-726ea76865ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.254519] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1222.254519] env[65121]: value = "task-5107537" [ 1222.254519] env[65121]: _type = "Task" [ 1222.254519] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.266132] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.391444] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.284s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1222.391858] env[65121]: DEBUG nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1222.685983] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107535, 'name': ReconfigVM_Task, 'duration_secs': 0.177229} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.686728] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487569b0-3ba8-43f4-baa4-999f574dca9b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.707738] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1222.708009] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1222.708169] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1222.708346] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1222.708523] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1222.708761] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1222.708819] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1222.708970] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1222.709085] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1222.709384] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1222.709445] env[65121]: DEBUG nova.virt.hardware [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1222.710782] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61d98a31-0894-4b29-bb83-64e5f397fcec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.718107] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1222.718107] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a4e7a3-5fb5-47eb-b799-243a45f603cc" [ 1222.718107] env[65121]: _type = "Task" [ 1222.718107] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.727345] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a4e7a3-5fb5-47eb-b799-243a45f603cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.765126] env[65121]: DEBUG oslo_vmware.api [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243441} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.765393] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1222.765574] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1222.765742] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1222.786837] env[65121]: INFO nova.scheduler.client.report [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted allocations for instance 999ab257-c116-4a73-9b87-b52aeb9a3bb3 [ 1222.896872] env[65121]: DEBUG nova.compute.utils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1222.898394] env[65121]: DEBUG nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1222.898597] env[65121]: DEBUG nova.network.neutron [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1222.898979] env[65121]: WARNING neutronclient.v2_0.client [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.899373] env[65121]: WARNING neutronclient.v2_0.client [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.899976] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.900330] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.945401] env[65121]: DEBUG nova.policy [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8931a266fddc4346b0c5243608a39bff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00383aa3355e438cb703c2b86c7917f2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1223.231664] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a4e7a3-5fb5-47eb-b799-243a45f603cc, 'name': SearchDatastore_Task, 'duration_secs': 0.011156} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.239067] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1223.239067] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2fc5a510-bbd5-4c60-8e5f-b8bca4f3a249 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.259106] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1223.259106] env[65121]: value = "task-5107538" [ 1223.259106] env[65121]: _type = "Task" [ 1223.259106] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.260848] env[65121]: DEBUG nova.network.neutron [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Successfully created port: 76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1223.274542] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107538, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.293053] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1223.293053] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1223.293053] env[65121]: DEBUG nova.objects.instance [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'resources' on Instance uuid 999ab257-c116-4a73-9b87-b52aeb9a3bb3 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1223.408286] env[65121]: DEBUG nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1223.776139] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107538, 'name': ReconfigVM_Task, 'duration_secs': 0.214189} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.776139] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1223.776139] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15703377-0e93-4846-b084-b55b88028cba {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.800140] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1223.800859] env[65121]: DEBUG nova.objects.instance [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'numa_topology' on Instance uuid 999ab257-c116-4a73-9b87-b52aeb9a3bb3 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1223.801994] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95599d0e-6340-46e2-ba81-facbb45996eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.818037] env[65121]: DEBUG nova.objects.base [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Object Instance<999ab257-c116-4a73-9b87-b52aeb9a3bb3> lazy-loaded attributes: resources,numa_topology {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1223.828282] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1223.828282] env[65121]: value = "task-5107539" [ 1223.828282] env[65121]: _type = "Task" [ 1223.828282] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.839659] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107539, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.915486] env[65121]: INFO nova.virt.block_device [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Booting with volume 47c28970-ebc3-443e-a72c-4ac58b364e40 at /dev/sda [ 1223.951344] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93fec9b4-7037-4c56-b5d1-f29b86325aac {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.962081] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba19fb4-dc87-453b-8cc0-f802b2d7ee5a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.988213] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d66046f-885d-44e2-b01d-143819045d41 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.003542] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36bf0a70-59fb-4e64-b39c-9d7f6c9c951d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.009339] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c36515-19d5-4668-927a-dab6000be6df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.017018] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8471f225-3167-43bf-9eaa-7464a3e6c233 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.054671] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2f42b8-9468-43e3-995b-b34a2629ad7c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.066390] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e1a883-7302-49d9-b0e5-fbe8c91e65b8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.085989] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97771652-3834-4af8-ae6f-38dcdaa4395e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.090017] env[65121]: DEBUG nova.compute.manager [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1224.090201] env[65121]: DEBUG nova.compute.manager [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing instance network info cache due to event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1224.090474] env[65121]: DEBUG oslo_concurrency.lockutils [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Acquiring lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.090620] env[65121]: DEBUG oslo_concurrency.lockutils [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Acquired lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.090793] env[65121]: DEBUG nova.network.neutron [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1224.100614] env[65121]: DEBUG nova.compute.provider_tree [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1224.102841] env[65121]: WARNING neutronclient.v2_0.client [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.103480] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.103944] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.113199] env[65121]: DEBUG nova.scheduler.client.report [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1224.120049] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cd281f-f520-4fbc-a76f-dea9ec853646 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.137052] env[65121]: DEBUG nova.virt.block_device [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating existing volume attachment record: ac4e4560-9535-4de3-8c34-40166f83f815 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1224.239619] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.240078] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.305758] env[65121]: WARNING neutronclient.v2_0.client [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.306737] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.307348] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.339067] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107539, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.375219] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.375910] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.394882] env[65121]: DEBUG nova.network.neutron [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updated VIF entry in instance network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1224.395265] env[65121]: DEBUG nova.network.neutron [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap91969a04-4a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1224.620859] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.328s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.841384] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107539, 'name': ReconfigVM_Task, 'duration_secs': 0.746133} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.841693] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Reconfigured VM instance instance-0000006c to attach disk [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5/f58bf796-69e7-4a61-9cea-78dff8d4fab5.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1224.842519] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74045729-df4c-4d06-bfac-de5a94ee2737 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.862269] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6213dde-6d3f-4821-8b53-14a072418838 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.885729] env[65121]: DEBUG nova.network.neutron [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Successfully updated port: 76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1224.888197] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af178ef5-d8b7-4240-a2db-48d91c7c486f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.890329] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.891111] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.891749] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.891749] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.891749] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.892107] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_power_states {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.910670] env[65121]: DEBUG oslo_concurrency.lockutils [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Releasing lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1224.910859] env[65121]: DEBUG nova.compute.manager [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Received event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1224.911046] env[65121]: DEBUG nova.compute.manager [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing instance network info cache due to event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1224.911263] env[65121]: DEBUG oslo_concurrency.lockutils [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Acquiring lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.911402] env[65121]: DEBUG oslo_concurrency.lockutils [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Acquired lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.915203] env[65121]: DEBUG nova.network.neutron [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1224.920203] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc99d468-ee5f-4919-b4b2-96cc8e21331c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.929203] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1224.930401] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c5bce0e-9b68-4d22-a217-03735323b729 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.933479] env[65121]: DEBUG nova.compute.manager [req-f58d980a-25d0-433f-898c-471018fd345a req-b9516bad-dae1-4e2a-b6fc-0633d4427fb4 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Received event network-vif-plugged-76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1224.933676] env[65121]: DEBUG oslo_concurrency.lockutils [req-f58d980a-25d0-433f-898c-471018fd345a req-b9516bad-dae1-4e2a-b6fc-0633d4427fb4 service nova] Acquiring lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.933868] env[65121]: DEBUG oslo_concurrency.lockutils [req-f58d980a-25d0-433f-898c-471018fd345a req-b9516bad-dae1-4e2a-b6fc-0633d4427fb4 service nova] Lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.934032] env[65121]: DEBUG oslo_concurrency.lockutils [req-f58d980a-25d0-433f-898c-471018fd345a req-b9516bad-dae1-4e2a-b6fc-0633d4427fb4 service nova] Lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.934190] env[65121]: DEBUG nova.compute.manager [req-f58d980a-25d0-433f-898c-471018fd345a req-b9516bad-dae1-4e2a-b6fc-0633d4427fb4 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] No waiting events found dispatching network-vif-plugged-76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1224.934343] env[65121]: WARNING nova.compute.manager [req-f58d980a-25d0-433f-898c-471018fd345a req-b9516bad-dae1-4e2a-b6fc-0633d4427fb4 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Received unexpected event network-vif-plugged-76e75737-0e97-46ba-89f7-b94bbcbff126 for instance with vm_state building and task_state block_device_mapping. [ 1224.941862] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1224.941862] env[65121]: value = "task-5107540" [ 1224.941862] env[65121]: _type = "Task" [ 1224.941862] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.952031] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107540, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.133702] env[65121]: DEBUG oslo_concurrency.lockutils [None req-1d9a6027-8248-4c23-a45e-59eeacfbd026 tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.883s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.391962] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.392171] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1225.392535] env[65121]: DEBUG nova.network.neutron [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1225.396383] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Getting list of instances from cluster (obj){ [ 1225.396383] env[65121]: value = "domain-c8" [ 1225.396383] env[65121]: _type = "ClusterComputeResource" [ 1225.396383] env[65121]: } {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1225.397498] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f91771b-a3ef-4723-9e21-0fc9c7285e9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.423153] env[65121]: WARNING neutronclient.v2_0.client [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1225.423854] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.424251] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.431493] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Got total of 10 instances {{(pid=65121) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1225.431691] env[65121]: WARNING nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] While synchronizing instance power states, found 11 instances in the database and 10 instances on the hypervisor. [ 1225.431837] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid 4a72fedd-b114-468e-8f34-0caec6ce73fa {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.432158] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid f58bf796-69e7-4a61-9cea-78dff8d4fab5 {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.432495] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid 3859dd02-9b52-4c71-a41f-c41f47ac1706 {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.432724] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid d49a301c-c0d5-476e-9f1f-b8f98cf6adb7 {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.432958] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid 138e48fe-0221-4173-801b-ee53b2bb98df {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.433217] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.433443] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid 992ba328-206a-4d47-a28b-f3ef9b20817f {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.433702] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.433959] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid e19d6d9b-d15d-4780-8564-f479573ce59b {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.434221] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid d34b8457-adc1-4d6b-befd-f6f1aff6d1fb {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.434556] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Triggering sync for uuid 91a36110-5e97-4553-90c3-3e4213e30c14 {{(pid=65121) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1225.436044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.436199] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.436426] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.436883] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.437165] env[65121]: INFO nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1225.437165] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.437491] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.437748] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.438122] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.438389] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.438775] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.439077] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.439500] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.439708] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.439988] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "992ba328-206a-4d47-a28b-f3ef9b20817f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.440197] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.440438] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.440615] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.440870] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "e19d6d9b-d15d-4780-8564-f479573ce59b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.441051] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.441310] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.441506] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.442704] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "91a36110-5e97-4553-90c3-3e4213e30c14" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.442704] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.442704] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1225.443348] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d441946-b3ce-4487-a952-7bbc84a3b9c2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.446832] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0a933e-3eeb-4a54-8c1f-b65807b0dcab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.450894] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f1c7c6-e1a2-4d15-b788-0cc5d5290dc4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.453938] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b8f270-c7e6-4194-ad0b-ab34b24629c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.457207] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd87232-f483-4750-a96b-dc2a34288d38 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.460243] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec92486-cc2d-479c-8e2f-71d5c521bf60 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.463324] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fb9607-80ec-4871-90de-84e16706b4a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.466882] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75060a0d-e1b0-4696-b48e-a2d075235527 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.470542] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef34667-660c-4450-a056-e694381b4126 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.473609] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.490070] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.490441] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.490606] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.490772] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1225.495494] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-005350c8-9c70-435d-9461-9f4c402a07b0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.542351] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107540, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.548407] env[65121]: WARNING urllib3.connectionpool [-] Connection pool is full, discarding connection: vc1.osci.c.eu-de-1.cloud.sap. Connection pool size: 10: queue.Full [ 1225.553758] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0932bf37-9ec7-46e1-bab9-a5fb1d719abc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.582750] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69094e9-c6f9-4f42-b244-aa8b5bafd31a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.593384] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5be41ff-2d28-41de-b116-49cb625cd182 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.626276] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178676MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1225.626455] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.626652] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.629694] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.630060] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.695679] env[65121]: WARNING neutronclient.v2_0.client [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1225.696453] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.696746] env[65121]: WARNING openstack [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.783252] env[65121]: DEBUG nova.network.neutron [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updated VIF entry in instance network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1225.783673] env[65121]: DEBUG nova.network.neutron [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1225.895742] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.896170] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.940305] env[65121]: DEBUG nova.network.neutron [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1225.962878] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.963298] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.983980] env[65121]: DEBUG oslo_vmware.api [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107540, 'name': PowerOnVM_Task, 'duration_secs': 0.673819} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.984304] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1226.022048] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.583s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.026320] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.590s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.029128] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.589s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.035862] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.595s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.036559] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.595s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.036789] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.595s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.049505] env[65121]: WARNING neutronclient.v2_0.client [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1226.050213] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.050676] env[65121]: WARNING openstack [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.066967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1226.067163] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.067379] env[65121]: INFO nova.compute.manager [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Unshelving [ 1226.068856] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.630s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.069237] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.632s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.069903] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.630s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.134828] env[65121]: DEBUG nova.network.neutron [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [{"id": "76e75737-0e97-46ba-89f7-b94bbcbff126", "address": "fa:16:3e:38:42:7e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e75737-0e", "ovs_interfaceid": "76e75737-0e97-46ba-89f7-b94bbcbff126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1226.239374] env[65121]: DEBUG nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1226.239948] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1226.240168] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1226.240326] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1226.240495] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1226.240650] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1226.240765] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1226.241101] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1226.241216] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1226.241359] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1226.242034] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1226.242034] env[65121]: DEBUG nova.virt.hardware [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1226.242606] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8012f351-0a39-4e39-8c51-da1111538b7f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.251664] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a071149d-e148-4c80-adc2-7b6fa6f6d053 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.286595] env[65121]: DEBUG oslo_concurrency.lockutils [req-d3ce575e-30f4-4906-a5bc-619feee3dab6 req-f19fb0cf-aa91-48c8-82d7-1f3b76eea8fa service nova] Releasing lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.438429] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1226.438906] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.637967] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.638388] env[65121]: DEBUG nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Instance network_info: |[{"id": "76e75737-0e97-46ba-89f7-b94bbcbff126", "address": "fa:16:3e:38:42:7e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e75737-0e", "ovs_interfaceid": "76e75737-0e97-46ba-89f7-b94bbcbff126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1226.639229] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:42:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76e75737-0e97-46ba-89f7-b94bbcbff126', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1226.646712] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1226.650722] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1226.652057] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c1910ed-7387-4b64-90c5-9ae87eba6ee8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.671317] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4a72fedd-b114-468e-8f34-0caec6ce73fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.671462] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 3859dd02-9b52-4c71-a41f-c41f47ac1706 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.671573] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance d49a301c-c0d5-476e-9f1f-b8f98cf6adb7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.671687] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 138e48fe-0221-4173-801b-ee53b2bb98df actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.671796] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.671905] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 992ba328-206a-4d47-a28b-f3ef9b20817f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.672023] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance f58bf796-69e7-4a61-9cea-78dff8d4fab5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.672134] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance db417aa7-d170-45bc-b2c4-7a97400de1d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.672240] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance e19d6d9b-d15d-4780-8564-f479573ce59b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.672345] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance d34b8457-adc1-4d6b-befd-f6f1aff6d1fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.672449] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 91a36110-5e97-4553-90c3-3e4213e30c14 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1226.675722] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1226.675722] env[65121]: value = "task-5107541" [ 1226.675722] env[65121]: _type = "Task" [ 1226.675722] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.684884] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107541, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.945395] env[65121]: DEBUG nova.compute.utils [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1226.964015] env[65121]: DEBUG nova.compute.manager [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Received event network-changed-76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1226.964102] env[65121]: DEBUG nova.compute.manager [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Refreshing instance network info cache due to event network-changed-76e75737-0e97-46ba-89f7-b94bbcbff126. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1226.964429] env[65121]: DEBUG oslo_concurrency.lockutils [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Acquiring lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.964429] env[65121]: DEBUG oslo_concurrency.lockutils [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Acquired lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1226.964515] env[65121]: DEBUG nova.network.neutron [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Refreshing network info cache for port 76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1226.995423] env[65121]: INFO nova.compute.manager [None req-520a2190-aee4-48b4-acf2-6caa312c1c96 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance to original state: 'active' [ 1227.090702] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.177646] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 999ab257-c116-4a73-9b87-b52aeb9a3bb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1227.177895] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1227.178070] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=100GB used_disk=10GB total_vcpus=48 used_vcpus=11 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '11', 'num_vm_active': '8', 'num_task_None': '9', 'num_os_type_None': '11', 'num_proj_00383aa3355e438cb703c2b86c7917f2': '3', 'io_workload': '1', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_72237a0d762645588c41231b0a34a796': '1', 'num_proj_a23c876483f34b1db8beec9a669f7822': '3', 'num_vm_rescued': '1', 'num_proj_f056059180af48bba4587006efb0b211': '1', 'num_proj_1ce7f6698e214d73ae43427601058af8': '1', 'num_proj_4467da3ed41245ddbc93fc865a8b7bdd': '1', 'num_proj_b07f57e89c774ef1bbaff256800ab80b': '1', 'num_vm_building': '1', 'num_task_block_device_mapping': '1'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1227.190320] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107541, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.341096] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4db4cc-881b-42ca-b6cd-c2d60aa90d67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.349881] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec78870-398b-4f34-99d4-5d445c7d1f56 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.384307] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe6d5ff-79ba-473e-97b4-ac9117cc9a95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.393046] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-851c32f7-e606-48af-945e-d4353b0c81c6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.406697] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1227.448496] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1227.467133] env[65121]: WARNING neutronclient.v2_0.client [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1227.468190] env[65121]: WARNING openstack [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1227.468582] env[65121]: WARNING openstack [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1227.564601] env[65121]: WARNING openstack [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1227.565013] env[65121]: WARNING openstack [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1227.625032] env[65121]: WARNING neutronclient.v2_0.client [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1227.625647] env[65121]: WARNING openstack [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1227.626049] env[65121]: WARNING openstack [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1227.696374] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107541, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.735990] env[65121]: DEBUG nova.network.neutron [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updated VIF entry in instance network info cache for port 76e75737-0e97-46ba-89f7-b94bbcbff126. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1227.736588] env[65121]: DEBUG nova.network.neutron [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [{"id": "76e75737-0e97-46ba-89f7-b94bbcbff126", "address": "fa:16:3e:38:42:7e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e75737-0e", "ovs_interfaceid": "76e75737-0e97-46ba-89f7-b94bbcbff126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1227.909546] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1228.152121] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1228.152572] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.152925] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1228.153253] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.153559] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1228.160037] env[65121]: INFO nova.compute.manager [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Terminating instance [ 1228.192924] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107541, 'name': CreateVM_Task, 'duration_secs': 1.417101} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.193163] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1228.193634] env[65121]: WARNING neutronclient.v2_0.client [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1228.193990] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'attachment_id': 'ac4e4560-9535-4de3-8c34-40166f83f815', 'device_type': None, 'delete_on_termination': True, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993593', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'name': 'volume-47c28970-ebc3-443e-a72c-4ac58b364e40', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '91a36110-5e97-4553-90c3-3e4213e30c14', 'attached_at': '', 'detached_at': '', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'serial': '47c28970-ebc3-443e-a72c-4ac58b364e40'}, 'mount_device': '/dev/sda', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65121) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1228.194207] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Root volume attach. Driver type: vmdk {{(pid=65121) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1228.195032] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f5112c-d277-4450-9d76-6621294e9083 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.204175] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0184fb-4bfc-4c09-b539-a4647b390ee7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.212448] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edf5af4-8a8a-4b84-b9f6-df1aa744d549 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.220082] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8a78d9ef-bcdb-4768-8834-38f4629993d9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.229234] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1228.229234] env[65121]: value = "task-5107542" [ 1228.229234] env[65121]: _type = "Task" [ 1228.229234] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.239066] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.239636] env[65121]: DEBUG oslo_concurrency.lockutils [req-df716bfb-f65d-4e40-b4a6-d7e685248c17 req-d85560f1-8a42-40ce-bffe-16b4f7397fc7 service nova] Releasing lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1228.415059] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1228.415312] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.789s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1228.415607] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.325s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.415805] env[65121]: DEBUG nova.objects.instance [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'pci_requests' on Instance uuid 999ab257-c116-4a73-9b87-b52aeb9a3bb3 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1228.524486] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1228.524792] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.525028] env[65121]: INFO nova.compute.manager [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Attaching volume 7c6ea767-b3df-4bc0-b083-3f50458d5082 to /dev/sdb [ 1228.568872] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de775b8-2374-421c-bcc3-2f5efd0bbb71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.580940] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a89a948-c1fd-483d-b263-193943c46133 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.603121] env[65121]: DEBUG nova.virt.block_device [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Updating existing volume attachment record: e41d752d-d71c-45a9-b151-8102f695d772 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1228.665590] env[65121]: DEBUG nova.compute.manager [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1228.665908] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1228.667213] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16be40ff-6d64-437c-91d0-a43176339388 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.677389] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1228.677714] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f754579c-1957-4585-a5bc-b47daa6ebe48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.688430] env[65121]: DEBUG oslo_vmware.api [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1228.688430] env[65121]: value = "task-5107543" [ 1228.688430] env[65121]: _type = "Task" [ 1228.688430] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.700217] env[65121]: DEBUG oslo_vmware.api [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107543, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.742482] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task} progress is 40%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.921125] env[65121]: DEBUG nova.objects.instance [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'numa_topology' on Instance uuid 999ab257-c116-4a73-9b87-b52aeb9a3bb3 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.200313] env[65121]: DEBUG oslo_vmware.api [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107543, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.245785] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.423800] env[65121]: INFO nova.compute.claims [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1229.703113] env[65121]: DEBUG oslo_vmware.api [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107543, 'name': PowerOffVM_Task, 'duration_secs': 0.679778} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.703113] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1229.703392] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1229.703434] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbce5db2-9b1e-46a3-97a8-6f26cb8f4cc1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.742691] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task} progress is 63%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.805081] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1229.805081] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1229.805503] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleting the datastore file [datastore1] f58bf796-69e7-4a61-9cea-78dff8d4fab5 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1229.806026] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d6f604f-e3cd-4f4d-8673-4d152ffe914c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.816224] env[65121]: DEBUG oslo_vmware.api [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1229.816224] env[65121]: value = "task-5107546" [ 1229.816224] env[65121]: _type = "Task" [ 1229.816224] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.829787] env[65121]: DEBUG oslo_vmware.api [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107546, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.246414] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task} progress is 76%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.330466] env[65121]: DEBUG oslo_vmware.api [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107546, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187383} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.330797] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1230.330987] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1230.331354] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1230.331593] env[65121]: INFO nova.compute.manager [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Took 1.67 seconds to destroy the instance on the hypervisor. [ 1230.331904] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1230.332183] env[65121]: DEBUG nova.compute.manager [-] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1230.332294] env[65121]: DEBUG nova.network.neutron [-] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1230.332750] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1230.333364] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1230.333660] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1230.375765] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1230.622813] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b5c410-cde2-44af-9d84-69a8003e95b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.633541] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702c46fe-4d36-4941-8ddf-a57ce1560870 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.671887] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ce2483-8275-467b-b515-e7ee01e80bf3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.682680] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174cb7f9-3d7d-4b8e-a047-fe97746cb0b9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.702939] env[65121]: DEBUG nova.compute.provider_tree [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1230.748888] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.931186] env[65121]: DEBUG nova.compute.manager [req-8ab30cac-487e-4d06-9bc0-780b12369799 req-d4b7ae47-cc56-4668-aec2-0e1ef4616dae service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Received event network-vif-deleted-99be1255-3e71-41e6-936a-acbf44bf9de9 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1230.931186] env[65121]: INFO nova.compute.manager [req-8ab30cac-487e-4d06-9bc0-780b12369799 req-d4b7ae47-cc56-4668-aec2-0e1ef4616dae service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Neutron deleted interface 99be1255-3e71-41e6-936a-acbf44bf9de9; detaching it from the instance and deleting it from the info cache [ 1230.931186] env[65121]: DEBUG nova.network.neutron [req-8ab30cac-487e-4d06-9bc0-780b12369799 req-d4b7ae47-cc56-4668-aec2-0e1ef4616dae service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1231.206894] env[65121]: DEBUG nova.scheduler.client.report [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1231.244193] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task} progress is 97%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.399255] env[65121]: DEBUG nova.network.neutron [-] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1231.437521] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a2edab2-fa8a-4910-abed-b19a92d6d30a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.448811] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5efb7b-0a6b-4c3b-8035-e1d7be1f808c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.492133] env[65121]: DEBUG nova.compute.manager [req-8ab30cac-487e-4d06-9bc0-780b12369799 req-d4b7ae47-cc56-4668-aec2-0e1ef4616dae service nova] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Detach interface failed, port_id=99be1255-3e71-41e6-936a-acbf44bf9de9, reason: Instance f58bf796-69e7-4a61-9cea-78dff8d4fab5 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1231.711853] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.296s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1231.712353] env[65121]: WARNING neutronclient.v2_0.client [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1231.743827] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107542, 'name': RelocateVM_Task, 'duration_secs': 3.474005} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.744865] env[65121]: INFO nova.network.neutron [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating port 91969a04-4a8e-4334-829f-3e3ae6828511 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1231.746968] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1231.747196] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993593', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'name': 'volume-47c28970-ebc3-443e-a72c-4ac58b364e40', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '91a36110-5e97-4553-90c3-3e4213e30c14', 'attached_at': '', 'detached_at': '', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'serial': '47c28970-ebc3-443e-a72c-4ac58b364e40'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1231.748223] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ceae585-5548-4105-a47a-a9a890420c13 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.764751] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88926563-b5de-4eb0-9cd1-d525b05525c5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.789776] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] volume-47c28970-ebc3-443e-a72c-4ac58b364e40/volume-47c28970-ebc3-443e-a72c-4ac58b364e40.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1231.790115] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-932443dd-f2c8-4728-83d7-cf72c8f7c015 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.811698] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1231.811698] env[65121]: value = "task-5107548" [ 1231.811698] env[65121]: _type = "Task" [ 1231.811698] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.820543] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107548, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.903682] env[65121]: INFO nova.compute.manager [-] [instance: f58bf796-69e7-4a61-9cea-78dff8d4fab5] Took 1.57 seconds to deallocate network for instance. [ 1232.007670] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1232.008121] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1232.324080] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.412790] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1232.413161] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1232.413334] env[65121]: DEBUG nova.objects.instance [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'resources' on Instance uuid f58bf796-69e7-4a61-9cea-78dff8d4fab5 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.512319] env[65121]: DEBUG nova.compute.utils [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1232.823478] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107548, 'name': ReconfigVM_Task, 'duration_secs': 0.88581} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.823841] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfigured VM instance instance-0000007b to attach disk [datastore1] volume-47c28970-ebc3-443e-a72c-4ac58b364e40/volume-47c28970-ebc3-443e-a72c-4ac58b364e40.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1232.828886] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bdb3b4d-d63f-49e0-8e2c-8d3a369cece7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.847535] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1232.847535] env[65121]: value = "task-5107549" [ 1232.847535] env[65121]: _type = "Task" [ 1232.847535] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.856758] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107549, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.014833] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.083361] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18dc3dc-4d48-4c13-b634-702eabf2a869 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.091374] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a889e090-1fcc-480c-98cd-f17daa736fbb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.124214] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f76b2b-b2f3-42f0-8fa3-f54d27e55a0c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.132739] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a29cd6-f62c-4c38-8bee-6c55ea7ecbb5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.147308] env[65121]: DEBUG nova.compute.provider_tree [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.156975] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1233.157218] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993596', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'name': 'volume-7c6ea767-b3df-4bc0-b083-3f50458d5082', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '138e48fe-0221-4173-801b-ee53b2bb98df', 'attached_at': '', 'detached_at': '', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'serial': '7c6ea767-b3df-4bc0-b083-3f50458d5082'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1233.158383] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8547d391-b380-4d7b-ba8a-d5277b2a0635 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.176752] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50c033d-33bd-4b37-9290-ee9679c57ca3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.203237] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] volume-7c6ea767-b3df-4bc0-b083-3f50458d5082/volume-7c6ea767-b3df-4bc0-b083-3f50458d5082.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1233.204990] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31ac4a1e-7077-4cd5-a5eb-bac17f47dfc5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.219440] env[65121]: DEBUG nova.compute.manager [req-348616b6-89a8-4c9d-a187-396c0b96974c req-f90be528-d2b8-4b65-862b-ad0258481440 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-vif-plugged-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1233.219440] env[65121]: DEBUG oslo_concurrency.lockutils [req-348616b6-89a8-4c9d-a187-396c0b96974c req-f90be528-d2b8-4b65-862b-ad0258481440 service nova] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.219924] env[65121]: DEBUG oslo_concurrency.lockutils [req-348616b6-89a8-4c9d-a187-396c0b96974c req-f90be528-d2b8-4b65-862b-ad0258481440 service nova] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.220120] env[65121]: DEBUG oslo_concurrency.lockutils [req-348616b6-89a8-4c9d-a187-396c0b96974c req-f90be528-d2b8-4b65-862b-ad0258481440 service nova] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.220289] env[65121]: DEBUG nova.compute.manager [req-348616b6-89a8-4c9d-a187-396c0b96974c req-f90be528-d2b8-4b65-862b-ad0258481440 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] No waiting events found dispatching network-vif-plugged-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1233.220482] env[65121]: WARNING nova.compute.manager [req-348616b6-89a8-4c9d-a187-396c0b96974c req-f90be528-d2b8-4b65-862b-ad0258481440 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received unexpected event network-vif-plugged-91969a04-4a8e-4334-829f-3e3ae6828511 for instance with vm_state shelved_offloaded and task_state spawning. [ 1233.228074] env[65121]: DEBUG oslo_vmware.api [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1233.228074] env[65121]: value = "task-5107550" [ 1233.228074] env[65121]: _type = "Task" [ 1233.228074] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.238306] env[65121]: DEBUG oslo_vmware.api [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107550, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.296866] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.297125] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1233.297333] env[65121]: DEBUG nova.network.neutron [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1233.358821] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107549, 'name': ReconfigVM_Task, 'duration_secs': 0.150043} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.359202] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993593', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'name': 'volume-47c28970-ebc3-443e-a72c-4ac58b364e40', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '91a36110-5e97-4553-90c3-3e4213e30c14', 'attached_at': '', 'detached_at': '', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'serial': '47c28970-ebc3-443e-a72c-4ac58b364e40'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1233.359845] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1846d379-a71a-4ee3-8ac6-12b192498307 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.366875] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1233.366875] env[65121]: value = "task-5107551" [ 1233.366875] env[65121]: _type = "Task" [ 1233.366875] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.375800] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107551, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.650861] env[65121]: DEBUG nova.scheduler.client.report [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1233.737933] env[65121]: DEBUG oslo_vmware.api [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107550, 'name': ReconfigVM_Task, 'duration_secs': 0.367065} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.738308] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Reconfigured VM instance instance-00000074 to attach disk [datastore1] volume-7c6ea767-b3df-4bc0-b083-3f50458d5082/volume-7c6ea767-b3df-4bc0-b083-3f50458d5082.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1233.743068] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39b79283-b2cb-4fa2-972c-8971dceb900a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.759050] env[65121]: DEBUG oslo_vmware.api [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1233.759050] env[65121]: value = "task-5107552" [ 1233.759050] env[65121]: _type = "Task" [ 1233.759050] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.767398] env[65121]: DEBUG oslo_vmware.api [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.799913] env[65121]: WARNING neutronclient.v2_0.client [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1233.800924] env[65121]: WARNING openstack [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1233.801459] env[65121]: WARNING openstack [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1233.876896] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107551, 'name': Rename_Task, 'duration_secs': 0.131522} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.877213] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1233.877499] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c0845dc-d6c6-4a98-a6dd-9caf19e5d058 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.884805] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1233.884805] env[65121]: value = "task-5107553" [ 1233.884805] env[65121]: _type = "Task" [ 1233.884805] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.897661] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.937877] env[65121]: WARNING openstack [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1233.938424] env[65121]: WARNING openstack [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1234.005556] env[65121]: WARNING neutronclient.v2_0.client [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1234.006328] env[65121]: WARNING openstack [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1234.006675] env[65121]: WARNING openstack [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1234.087326] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1234.087326] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1234.087797] env[65121]: INFO nova.compute.manager [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Attaching volume 5d4df019-195e-4afc-bd9f-7e8d84c011f2 to /dev/sdb [ 1234.106096] env[65121]: DEBUG nova.network.neutron [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91969a04-4a", "ovs_interfaceid": "91969a04-4a8e-4334-829f-3e3ae6828511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1234.127916] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3c729a-8560-4bda-bb51-fedceab0ed22 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.137069] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa5f1f6-545d-4d27-947a-00eb6275200e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.152897] env[65121]: DEBUG nova.virt.block_device [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updating existing volume attachment record: 2de744d6-a941-4124-9d50-7a917b526062 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1234.156032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.185140] env[65121]: INFO nova.scheduler.client.report [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted allocations for instance f58bf796-69e7-4a61-9cea-78dff8d4fab5 [ 1234.270248] env[65121]: DEBUG oslo_vmware.api [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107552, 'name': ReconfigVM_Task, 'duration_secs': 0.153238} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.270600] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993596', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'name': 'volume-7c6ea767-b3df-4bc0-b083-3f50458d5082', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '138e48fe-0221-4173-801b-ee53b2bb98df', 'attached_at': '', 'detached_at': '', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'serial': '7c6ea767-b3df-4bc0-b083-3f50458d5082'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1234.395493] env[65121]: DEBUG oslo_vmware.api [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107553, 'name': PowerOnVM_Task, 'duration_secs': 0.489422} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.395770] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1234.396020] env[65121]: INFO nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Took 8.16 seconds to spawn the instance on the hypervisor. [ 1234.396281] env[65121]: DEBUG nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1234.397105] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7379768e-27d3-493e-9e2a-33a1a4e9ee42 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.610059] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1234.638233] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c48878b022239f7ca7f7ef434148e86d',container_format='bare',created_at=2025-12-12T14:30:31Z,direct_url=,disk_format='vmdk',id=ceda632d-08b0-45f0-ae23-8456ddb4144e,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1166068688-shelved',owner='ad1aea30d62c45e193c5a54c429ce7ba',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2025-12-12T14:30:47Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1234.638490] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1234.638689] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1234.638837] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1234.638980] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1234.639130] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1234.639328] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1234.639505] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1234.639674] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1234.639829] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1234.639994] env[65121]: DEBUG nova.virt.hardware [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1234.640872] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55376b0c-3d64-416d-a38d-f14bb52982a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.650334] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fec9d48-789c-4928-84fc-862fcea9f349 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.665812] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:16:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a64108f9-df0a-4feb-bbb5-97f5841c356c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91969a04-4a8e-4334-829f-3e3ae6828511', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1234.673454] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1234.673737] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1234.673961] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd7fbf38-c3cb-4ba3-8c2a-73af66b5151d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.693814] env[65121]: DEBUG oslo_concurrency.lockutils [None req-c4f11228-e416-443b-8e5d-bf062338f0e4 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "f58bf796-69e7-4a61-9cea-78dff8d4fab5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.541s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.696084] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1234.696084] env[65121]: value = "task-5107555" [ 1234.696084] env[65121]: _type = "Task" [ 1234.696084] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.705300] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107555, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.916894] env[65121]: INFO nova.compute.manager [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Took 14.83 seconds to build instance. [ 1235.207228] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107555, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.270031] env[65121]: DEBUG nova.compute.manager [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1235.270031] env[65121]: DEBUG nova.compute.manager [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing instance network info cache due to event network-changed-91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1235.270031] env[65121]: DEBUG oslo_concurrency.lockutils [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Acquiring lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.270226] env[65121]: DEBUG oslo_concurrency.lockutils [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Acquired lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1235.270418] env[65121]: DEBUG nova.network.neutron [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Refreshing network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1235.311749] env[65121]: DEBUG nova.objects.instance [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'flavor' on Instance uuid 138e48fe-0221-4173-801b-ee53b2bb98df {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.419447] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5a3ff489-01b1-43c9-8e55-e42af7f2c7e4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.340s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1235.419648] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.977s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.420783] env[65121]: INFO nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] During sync_power_state the instance has a pending task (block_device_mapping). Skip. [ 1235.420783] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1235.477352] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "b669fb5d-a68f-4390-b903-80879c239b4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.478931] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.707535] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107555, 'name': CreateVM_Task, 'duration_secs': 0.561461} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.707535] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1235.708021] env[65121]: WARNING neutronclient.v2_0.client [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.708243] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.708393] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1235.708851] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1235.709040] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e84bd30a-c597-4f70-8a58-ae10b7c532c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.714381] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1235.714381] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524054ac-ee05-6ccb-9a44-033a94fafbdb" [ 1235.714381] env[65121]: _type = "Task" [ 1235.714381] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.725975] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524054ac-ee05-6ccb-9a44-033a94fafbdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.773958] env[65121]: WARNING neutronclient.v2_0.client [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.774729] env[65121]: WARNING openstack [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1235.775192] env[65121]: WARNING openstack [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1235.818408] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f835690d-c70f-418c-b35a-3f53f7ef515b tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.294s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1235.896484] env[65121]: WARNING openstack [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1235.896942] env[65121]: WARNING openstack [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1235.971101] env[65121]: WARNING neutronclient.v2_0.client [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1235.971101] env[65121]: WARNING openstack [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1235.971101] env[65121]: WARNING openstack [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1235.980204] env[65121]: DEBUG nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1236.070349] env[65121]: DEBUG nova.network.neutron [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updated VIF entry in instance network info cache for port 91969a04-4a8e-4334-829f-3e3ae6828511. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1236.070726] env[65121]: DEBUG nova.network.neutron [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [{"id": "91969a04-4a8e-4334-829f-3e3ae6828511", "address": "fa:16:3e:e2:16:94", "network": {"id": "00f8e839-d398-4073-bd12-2662f1923643", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1600156853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad1aea30d62c45e193c5a54c429ce7ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91969a04-4a", "ovs_interfaceid": "91969a04-4a8e-4334-829f-3e3ae6828511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.085216] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1236.085459] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.117643] env[65121]: DEBUG nova.compute.manager [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Received event network-changed-4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1236.117643] env[65121]: DEBUG nova.compute.manager [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Refreshing instance network info cache due to event network-changed-4304bec2-f452-4be6-aeda-3ad8354a4547. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1236.117643] env[65121]: DEBUG oslo_concurrency.lockutils [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Acquiring lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.117643] env[65121]: DEBUG oslo_concurrency.lockutils [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Acquired lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1236.117957] env[65121]: DEBUG nova.network.neutron [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Refreshing network info cache for port 4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1236.225223] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1236.225481] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Processing image ceda632d-08b0-45f0-ae23-8456ddb4144e {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1236.225721] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e/ceda632d-08b0-45f0-ae23-8456ddb4144e.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.225861] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e/ceda632d-08b0-45f0-ae23-8456ddb4144e.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1236.226044] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1236.226301] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2b4d0a3-2ad2-41fa-b733-238e0649dc23 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.236537] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1236.236707] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1236.237464] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97ed8d65-0426-4d7f-8086-b833fd6ccc87 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.243486] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1236.243486] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522df0fa-e356-9bcd-32be-853879b7fb6d" [ 1236.243486] env[65121]: _type = "Task" [ 1236.243486] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.252546] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522df0fa-e356-9bcd-32be-853879b7fb6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.505996] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1236.506336] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.508247] env[65121]: INFO nova.compute.claims [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1236.574110] env[65121]: DEBUG oslo_concurrency.lockutils [req-ac24b3f3-d1f3-4c43-b127-a2acc31fa9b6 req-71f796fc-9f42-4ac0-9d58-7ad0a1e35823 service nova] Releasing lock "refresh_cache-999ab257-c116-4a73-9b87-b52aeb9a3bb3" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1236.589736] env[65121]: INFO nova.compute.manager [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Detaching volume 7c6ea767-b3df-4bc0-b083-3f50458d5082 [ 1236.621478] env[65121]: WARNING neutronclient.v2_0.client [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.623291] env[65121]: WARNING openstack [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.623959] env[65121]: WARNING openstack [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.638584] env[65121]: INFO nova.virt.block_device [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Attempting to driver detach volume 7c6ea767-b3df-4bc0-b083-3f50458d5082 from mountpoint /dev/sdb [ 1236.638930] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1236.639252] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993596', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'name': 'volume-7c6ea767-b3df-4bc0-b083-3f50458d5082', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '138e48fe-0221-4173-801b-ee53b2bb98df', 'attached_at': '', 'detached_at': '', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'serial': '7c6ea767-b3df-4bc0-b083-3f50458d5082'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1236.640968] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87285987-0d38-45ba-9471-7deaafa025ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.666352] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96349e68-f134-4295-8dd7-3781e9224bfb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.674656] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e57ced-f603-42e3-bf0a-c463357cd7bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.698211] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48416cf4-112e-4968-b119-866a5aeade0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.715451] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The volume has not been displaced from its original location: [datastore1] volume-7c6ea767-b3df-4bc0-b083-3f50458d5082/volume-7c6ea767-b3df-4bc0-b083-3f50458d5082.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1236.721280] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Reconfiguring VM instance instance-00000074 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1236.721711] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e81757b7-ffa2-4da5-a4c9-1114cfd05f46 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.750061] env[65121]: DEBUG oslo_vmware.api [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1236.750061] env[65121]: value = "task-5107557" [ 1236.750061] env[65121]: _type = "Task" [ 1236.750061] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.760103] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Preparing fetch location {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1236.760103] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Fetch image to [datastore1] OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403/OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403.vmdk {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1236.760103] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Downloading stream optimized image ceda632d-08b0-45f0-ae23-8456ddb4144e to [datastore1] OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403/OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403.vmdk on the data store datastore1 as vApp {{(pid=65121) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1236.760103] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Downloading image file data ceda632d-08b0-45f0-ae23-8456ddb4144e to the ESX as VM named 'OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403' {{(pid=65121) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1236.766788] env[65121]: WARNING openstack [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.767189] env[65121]: WARNING openstack [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.774952] env[65121]: DEBUG oslo_vmware.api [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107557, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.851317] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1236.851317] env[65121]: value = "resgroup-9" [ 1236.851317] env[65121]: _type = "ResourcePool" [ 1236.851317] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1236.851668] env[65121]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b148e671-0108-459e-97ff-ff153878ab4d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.868960] env[65121]: WARNING neutronclient.v2_0.client [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.869672] env[65121]: WARNING openstack [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.870024] env[65121]: WARNING openstack [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.884760] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease: (returnval){ [ 1236.884760] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525c3460-4f51-1819-814d-3459117edcab" [ 1236.884760] env[65121]: _type = "HttpNfcLease" [ 1236.884760] env[65121]: } obtained for vApp import into resource pool (val){ [ 1236.884760] env[65121]: value = "resgroup-9" [ 1236.884760] env[65121]: _type = "ResourcePool" [ 1236.884760] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1236.885250] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the lease: (returnval){ [ 1236.885250] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525c3460-4f51-1819-814d-3459117edcab" [ 1236.885250] env[65121]: _type = "HttpNfcLease" [ 1236.885250] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1236.892877] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1236.892877] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525c3460-4f51-1819-814d-3459117edcab" [ 1236.892877] env[65121]: _type = "HttpNfcLease" [ 1236.892877] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1236.969185] env[65121]: DEBUG nova.network.neutron [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updated VIF entry in instance network info cache for port 4304bec2-f452-4be6-aeda-3ad8354a4547. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1236.969608] env[65121]: DEBUG nova.network.neutron [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updating instance_info_cache with network_info: [{"id": "4304bec2-f452-4be6-aeda-3ad8354a4547", "address": "fa:16:3e:71:b3:4e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4304bec2-f4", "ovs_interfaceid": "4304bec2-f452-4be6-aeda-3ad8354a4547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1237.262093] env[65121]: DEBUG oslo_vmware.api [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107557, 'name': ReconfigVM_Task, 'duration_secs': 0.242977} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.262407] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Reconfigured VM instance instance-00000074 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1237.268155] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45ee5f9b-c3b5-46c1-8bd3-484abfee6749 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.284647] env[65121]: DEBUG oslo_vmware.api [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1237.284647] env[65121]: value = "task-5107559" [ 1237.284647] env[65121]: _type = "Task" [ 1237.284647] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.293201] env[65121]: DEBUG oslo_vmware.api [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107559, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.395664] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1237.395664] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525c3460-4f51-1819-814d-3459117edcab" [ 1237.395664] env[65121]: _type = "HttpNfcLease" [ 1237.395664] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1237.472762] env[65121]: DEBUG oslo_concurrency.lockutils [req-36fbe586-b4fe-4e45-9240-bd147f7fa676 req-7fd66927-37f5-490d-b3f3-8d591f79f62f service nova] Releasing lock "refresh_cache-4a72fedd-b114-468e-8f34-0caec6ce73fa" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1237.689454] env[65121]: DEBUG nova.compute.manager [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Stashing vm_state: active {{(pid=65121) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1237.709060] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1482645-874e-4f6e-9476-1f5457a27f3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.717266] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21039b2-cf84-4470-a61f-a6250c931859 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.752605] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f9e0a2-c469-4821-82cc-f50155e3efc1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.762288] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b89b2a3-ce76-419c-aafd-6d46a37b782d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.777015] env[65121]: DEBUG nova.compute.provider_tree [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.794799] env[65121]: DEBUG oslo_vmware.api [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107559, 'name': ReconfigVM_Task, 'duration_secs': 0.164752} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.795136] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993596', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'name': 'volume-7c6ea767-b3df-4bc0-b083-3f50458d5082', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '138e48fe-0221-4173-801b-ee53b2bb98df', 'attached_at': '', 'detached_at': '', 'volume_id': '7c6ea767-b3df-4bc0-b083-3f50458d5082', 'serial': '7c6ea767-b3df-4bc0-b083-3f50458d5082'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1237.894721] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1237.894721] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525c3460-4f51-1819-814d-3459117edcab" [ 1237.894721] env[65121]: _type = "HttpNfcLease" [ 1237.894721] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1237.895154] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1237.895154] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525c3460-4f51-1819-814d-3459117edcab" [ 1237.895154] env[65121]: _type = "HttpNfcLease" [ 1237.895154] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1237.895722] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef77eda2-06f0-4a90-bb7c-ac4ae460cc02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.904056] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c14f06-32fb-3de0-3b75-26a7aa29c2f0/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1237.904329] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c14f06-32fb-3de0-3b75-26a7aa29c2f0/disk-0.vmdk. {{(pid=65121) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1237.966083] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fb3b198f-3137-46ed-a652-bd91623bbbdb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.147916] env[65121]: DEBUG nova.compute.manager [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Received event network-changed-76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1238.148222] env[65121]: DEBUG nova.compute.manager [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Refreshing instance network info cache due to event network-changed-76e75737-0e97-46ba-89f7-b94bbcbff126. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1238.148512] env[65121]: DEBUG oslo_concurrency.lockutils [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Acquiring lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.148733] env[65121]: DEBUG oslo_concurrency.lockutils [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Acquired lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1238.149146] env[65121]: DEBUG nova.network.neutron [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Refreshing network info cache for port 76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1238.208495] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1238.279904] env[65121]: DEBUG nova.scheduler.client.report [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1238.337442] env[65121]: DEBUG nova.objects.instance [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'flavor' on Instance uuid 138e48fe-0221-4173-801b-ee53b2bb98df {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.652060] env[65121]: WARNING neutronclient.v2_0.client [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1238.652804] env[65121]: WARNING openstack [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.653209] env[65121]: WARNING openstack [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.704033] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1238.704274] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993597', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'name': 'volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'serial': '5d4df019-195e-4afc-bd9f-7e8d84c011f2'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1238.705252] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5907804-5a12-4676-8350-20623cec5cf8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.722373] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43715779-d495-4638-bcbb-d40be53b0bf2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.756593] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2/volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1238.759775] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca4392b5-1442-4fe1-9ccb-21f13316077d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.781139] env[65121]: WARNING openstack [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.781486] env[65121]: WARNING openstack [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.788929] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.789502] env[65121]: DEBUG nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1238.795332] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.587s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1238.796873] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1238.796873] env[65121]: value = "task-5107560" [ 1238.796873] env[65121]: _type = "Task" [ 1238.796873] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.812782] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107560, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.878306] env[65121]: WARNING neutronclient.v2_0.client [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1238.878974] env[65121]: WARNING openstack [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1238.879352] env[65121]: WARNING openstack [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1238.975722] env[65121]: DEBUG nova.network.neutron [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updated VIF entry in instance network info cache for port 76e75737-0e97-46ba-89f7-b94bbcbff126. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1238.976194] env[65121]: DEBUG nova.network.neutron [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [{"id": "76e75737-0e97-46ba-89f7-b94bbcbff126", "address": "fa:16:3e:38:42:7e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e75737-0e", "ovs_interfaceid": "76e75737-0e97-46ba-89f7-b94bbcbff126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1239.300541] env[65121]: DEBUG nova.compute.utils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1239.304851] env[65121]: INFO nova.compute.claims [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1239.308123] env[65121]: DEBUG nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1239.308883] env[65121]: DEBUG nova.network.neutron [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1239.308883] env[65121]: WARNING neutronclient.v2_0.client [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1239.309083] env[65121]: WARNING neutronclient.v2_0.client [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1239.309772] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.310114] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.333882] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107560, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.351719] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b3e0f525-ee6f-4a15-8685-6358435d9340 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.266s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1239.358322] env[65121]: DEBUG nova.policy [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af00bd582c0843949491bbcecbfcd2dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72237a0d762645588c41231b0a34a796', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1239.479157] env[65121]: DEBUG oslo_concurrency.lockutils [req-ba9e8815-568e-478d-bf17-9e4983e5df3b req-47ae7b5d-6601-4207-9288-731c3687f612 service nova] Releasing lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1239.717774] env[65121]: DEBUG nova.network.neutron [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Successfully created port: da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1239.804125] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Completed reading data from the image iterator. {{(pid=65121) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1239.804494] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c14f06-32fb-3de0-3b75-26a7aa29c2f0/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1239.805727] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c98cd6-651e-46dd-a26c-f4fbd1c1bddd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.820913] env[65121]: DEBUG nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1239.825471] env[65121]: INFO nova.compute.resource_tracker [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating resource usage from migration 37985e1d-7fb7-436b-9d96-696e712f7a17 [ 1239.828169] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c14f06-32fb-3de0-3b75-26a7aa29c2f0/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1239.828326] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c14f06-32fb-3de0-3b75-26a7aa29c2f0/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1239.832670] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ae7d3d9d-badd-4944-95b5-9db96ded1f8b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.834424] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107560, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.026113] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac794371-d2bc-4d40-a65d-b5a01ef36946 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.034919] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021c036a-1be7-4440-bf0a-0fe050c8627c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.067830] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2255b9ac-c40b-4df9-96ba-4bbdb8e19885 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.076622] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ab1f66-96a1-4f31-87b6-7abe425711b7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.090446] env[65121]: DEBUG nova.compute.provider_tree [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1240.316822] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107560, 'name': ReconfigVM_Task, 'duration_secs': 1.409415} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.317180] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfigured VM instance instance-00000075 to attach disk [datastore1] volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2/volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1240.322571] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fea40aa-0fdd-43c0-8977-e29e45ab1a38 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.342888] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1240.342888] env[65121]: value = "task-5107561" [ 1240.342888] env[65121]: _type = "Task" [ 1240.342888] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.353091] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107561, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.402089] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1240.402364] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.402567] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "138e48fe-0221-4173-801b-ee53b2bb98df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1240.402740] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.402893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.405508] env[65121]: INFO nova.compute.manager [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Terminating instance [ 1240.539038] env[65121]: DEBUG oslo_vmware.rw_handles [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c14f06-32fb-3de0-3b75-26a7aa29c2f0/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1240.539279] env[65121]: INFO nova.virt.vmwareapi.images [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Downloaded image file data ceda632d-08b0-45f0-ae23-8456ddb4144e [ 1240.540342] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b420c64a-d07a-4511-80f1-e2086565383b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.558790] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7fc88643-21ed-490b-a2e3-01a2b663baaa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.593886] env[65121]: DEBUG nova.scheduler.client.report [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1240.658035] env[65121]: INFO nova.virt.vmwareapi.images [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] The imported VM was unregistered [ 1240.660308] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Caching image {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1240.660905] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Creating directory with path [datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1240.661381] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4279ad83-23dc-4b53-a673-cda1e410f847 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.694772] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Created directory with path [datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1240.694772] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403/OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403.vmdk to [datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e/ceda632d-08b0-45f0-ae23-8456ddb4144e.vmdk. {{(pid=65121) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1240.695158] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7920f1dc-2dec-43d0-829b-4c74431b1f07 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.704326] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1240.704326] env[65121]: value = "task-5107563" [ 1240.704326] env[65121]: _type = "Task" [ 1240.704326] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.714820] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.838058] env[65121]: DEBUG nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1240.853054] env[65121]: DEBUG oslo_vmware.api [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107561, 'name': ReconfigVM_Task, 'duration_secs': 0.313039} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.853054] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993597', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'name': 'volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'serial': '5d4df019-195e-4afc-bd9f-7e8d84c011f2'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1240.871458] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1240.871790] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1240.871999] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1240.872267] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1240.872466] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1240.872704] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1240.872990] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1240.873240] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1240.873478] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1240.873705] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1240.873939] env[65121]: DEBUG nova.virt.hardware [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1240.875591] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86579288-eb29-4711-b8f6-930d2c025b77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.885077] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0757f62a-e91d-4f05-a277-36b3871a72c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.909535] env[65121]: DEBUG nova.compute.manager [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1240.909797] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1240.910942] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9869918a-50f8-4a32-b0a0-6df0b0241216 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.918324] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1240.918569] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db5aee9c-1da0-44f9-a166-5d199faaab67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.925501] env[65121]: DEBUG oslo_vmware.api [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1240.925501] env[65121]: value = "task-5107564" [ 1240.925501] env[65121]: _type = "Task" [ 1240.925501] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.934428] env[65121]: DEBUG oslo_vmware.api [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.099681] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.304s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1241.099895] env[65121]: INFO nova.compute.manager [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Migrating [ 1241.215178] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.219303] env[65121]: DEBUG nova.compute.manager [req-d95bfe80-888c-4cdd-a2ad-01dfdf0947b5 req-38f76c2a-01f3-4ffe-b6e2-1d0191c2d8ec service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Received event network-vif-plugged-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1241.219391] env[65121]: DEBUG oslo_concurrency.lockutils [req-d95bfe80-888c-4cdd-a2ad-01dfdf0947b5 req-38f76c2a-01f3-4ffe-b6e2-1d0191c2d8ec service nova] Acquiring lock "b669fb5d-a68f-4390-b903-80879c239b4e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1241.219664] env[65121]: DEBUG oslo_concurrency.lockutils [req-d95bfe80-888c-4cdd-a2ad-01dfdf0947b5 req-38f76c2a-01f3-4ffe-b6e2-1d0191c2d8ec service nova] Lock "b669fb5d-a68f-4390-b903-80879c239b4e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1241.219830] env[65121]: DEBUG oslo_concurrency.lockutils [req-d95bfe80-888c-4cdd-a2ad-01dfdf0947b5 req-38f76c2a-01f3-4ffe-b6e2-1d0191c2d8ec service nova] Lock "b669fb5d-a68f-4390-b903-80879c239b4e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1241.219986] env[65121]: DEBUG nova.compute.manager [req-d95bfe80-888c-4cdd-a2ad-01dfdf0947b5 req-38f76c2a-01f3-4ffe-b6e2-1d0191c2d8ec service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] No waiting events found dispatching network-vif-plugged-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1241.220156] env[65121]: WARNING nova.compute.manager [req-d95bfe80-888c-4cdd-a2ad-01dfdf0947b5 req-38f76c2a-01f3-4ffe-b6e2-1d0191c2d8ec service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Received unexpected event network-vif-plugged-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 for instance with vm_state building and task_state spawning. [ 1241.289797] env[65121]: DEBUG nova.network.neutron [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Successfully updated port: da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1241.436529] env[65121]: DEBUG oslo_vmware.api [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107564, 'name': PowerOffVM_Task, 'duration_secs': 0.217582} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.436853] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1241.436990] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1241.437305] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf06c42a-17b6-44b5-817b-fe077db70cab {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.502842] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1241.503113] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1241.503258] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleting the datastore file [datastore1] 138e48fe-0221-4173-801b-ee53b2bb98df {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1241.503526] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b64527a-e69c-465c-9b92-7b0c98099f82 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.514415] env[65121]: DEBUG oslo_vmware.api [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1241.514415] env[65121]: value = "task-5107566" [ 1241.514415] env[65121]: _type = "Task" [ 1241.514415] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.523275] env[65121]: DEBUG oslo_vmware.api [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.619808] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.620161] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1241.620407] env[65121]: DEBUG nova.network.neutron [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1241.716718] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.792990] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.793264] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1241.793482] env[65121]: DEBUG nova.network.neutron [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1241.889885] env[65121]: DEBUG nova.objects.instance [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1242.026356] env[65121]: DEBUG oslo_vmware.api [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.367824} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.026659] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1242.026810] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1242.026980] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1242.027486] env[65121]: INFO nova.compute.manager [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1242.027819] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1242.028036] env[65121]: DEBUG nova.compute.manager [-] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1242.028137] env[65121]: DEBUG nova.network.neutron [-] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1242.028432] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1242.029036] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.029559] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.080615] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1242.123719] env[65121]: WARNING neutronclient.v2_0.client [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1242.124462] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.124826] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.223552] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.264950] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.265548] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.296843] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.297409] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.399669] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35a776db-a671-4513-a008-5202c63d4036 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.312s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.415430] env[65121]: DEBUG nova.network.neutron [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1242.485141] env[65121]: WARNING neutronclient.v2_0.client [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1242.485676] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.486040] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.502906] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.503271] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.643294] env[65121]: DEBUG nova.network.neutron [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [{"id": "76e75737-0e97-46ba-89f7-b94bbcbff126", "address": "fa:16:3e:38:42:7e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e75737-0e", "ovs_interfaceid": "76e75737-0e97-46ba-89f7-b94bbcbff126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1242.664853] env[65121]: WARNING neutronclient.v2_0.client [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1242.665551] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1242.666012] env[65121]: WARNING openstack [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1242.728555] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.760110] env[65121]: DEBUG nova.network.neutron [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updating instance_info_cache with network_info: [{"id": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "address": "fa:16:3e:2b:98:34", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d86d6-62", "ovs_interfaceid": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1243.111340] env[65121]: DEBUG nova.network.neutron [-] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1243.146478] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1243.226382] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.258233] env[65121]: DEBUG nova.compute.manager [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Received event network-changed-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1243.258387] env[65121]: DEBUG nova.compute.manager [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Refreshing instance network info cache due to event network-changed-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1243.258546] env[65121]: DEBUG oslo_concurrency.lockutils [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Acquiring lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.262940] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1243.263295] env[65121]: DEBUG nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Instance network_info: |[{"id": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "address": "fa:16:3e:2b:98:34", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d86d6-62", "ovs_interfaceid": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1243.263614] env[65121]: DEBUG oslo_concurrency.lockutils [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Acquired lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.263783] env[65121]: DEBUG nova.network.neutron [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Refreshing network info cache for port da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1243.265087] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:98:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6966f473-59ac-49bb-9b7a-22c61f4e61e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1243.272837] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1243.274177] env[65121]: WARNING neutronclient.v2_0.client [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.274835] env[65121]: WARNING openstack [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.275208] env[65121]: WARNING openstack [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.283379] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1243.284164] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-254cd2ff-afaa-46a8-8cd6-b026897ed968 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.306719] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1243.306719] env[65121]: value = "task-5107567" [ 1243.306719] env[65121]: _type = "Task" [ 1243.306719] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.318928] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107567, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.374268] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1243.374521] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1243.395611] env[65121]: WARNING openstack [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.396421] env[65121]: WARNING openstack [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.469619] env[65121]: WARNING neutronclient.v2_0.client [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.470431] env[65121]: WARNING openstack [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.470909] env[65121]: WARNING openstack [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.564298] env[65121]: DEBUG nova.network.neutron [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updated VIF entry in instance network info cache for port da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1243.564683] env[65121]: DEBUG nova.network.neutron [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updating instance_info_cache with network_info: [{"id": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "address": "fa:16:3e:2b:98:34", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d86d6-62", "ovs_interfaceid": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1243.615069] env[65121]: INFO nova.compute.manager [-] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Took 1.59 seconds to deallocate network for instance. [ 1243.725580] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.818783] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107567, 'name': CreateVM_Task, 'duration_secs': 0.410382} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.818783] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1243.819029] env[65121]: WARNING neutronclient.v2_0.client [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.819456] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.819662] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.819996] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1243.820284] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c2355f0-885d-4482-8ba8-dca4c30cd589 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.825996] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1243.825996] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]525514b7-926d-6c95-a2ec-f32523db8e03" [ 1243.825996] env[65121]: _type = "Task" [ 1243.825996] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.837724] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525514b7-926d-6c95-a2ec-f32523db8e03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.878907] env[65121]: DEBUG nova.compute.utils [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1244.067967] env[65121]: DEBUG oslo_concurrency.lockutils [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] Releasing lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.068273] env[65121]: DEBUG nova.compute.manager [req-19fcd6af-672a-4aa6-8446-b7d028f23d04 req-371a766f-d3a2-4c54-8ce9-db2efdd190be service nova] [instance: 138e48fe-0221-4173-801b-ee53b2bb98df] Received event network-vif-deleted-3987861c-5f58-4f3f-868e-31452ab11d4d {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1244.121297] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1244.121740] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1244.121861] env[65121]: DEBUG nova.objects.instance [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'resources' on Instance uuid 138e48fe-0221-4173-801b-ee53b2bb98df {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.225177] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107563, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.287297} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.225464] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403/OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403.vmdk to [datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e/ceda632d-08b0-45f0-ae23-8456ddb4144e.vmdk. [ 1244.225681] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Cleaning up location [datastore1] OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1244.225845] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_0eec652f-8372-4b53-9fdb-af64fe5d1403 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1244.226128] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e01b5aa9-a845-48b3-b919-60451dedb3fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.233254] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1244.233254] env[65121]: value = "task-5107568" [ 1244.233254] env[65121]: _type = "Task" [ 1244.233254] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.241874] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107568, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.337270] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]525514b7-926d-6c95-a2ec-f32523db8e03, 'name': SearchDatastore_Task, 'duration_secs': 0.088218} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.337573] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.337789] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1244.338032] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.338166] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1244.338337] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1244.338606] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f1ffcfe-8d23-4388-a248-f9dab392a6c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.351785] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1244.352085] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1244.352967] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e4d6d65-2e8e-43e2-85b5-fec9c67deaea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.359158] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1244.359158] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5223e974-e608-de00-a80d-fb21c10467e8" [ 1244.359158] env[65121]: _type = "Task" [ 1244.359158] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.367542] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5223e974-e608-de00-a80d-fb21c10467e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.382663] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1244.662771] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0937592e-87a9-4fbf-97fa-76b88d6b6d99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.682060] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance '91a36110-5e97-4553-90c3-3e4213e30c14' progress to 0 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1244.743817] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107568, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195135} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.744076] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1244.744236] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e/ceda632d-08b0-45f0-ae23-8456ddb4144e.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.744463] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e/ceda632d-08b0-45f0-ae23-8456ddb4144e.vmdk to [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1244.744707] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a105c202-3e47-42e6-bf6e-de830b0b12d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.754229] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1244.754229] env[65121]: value = "task-5107569" [ 1244.754229] env[65121]: _type = "Task" [ 1244.754229] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.762649] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107569, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.809999] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae8fa70-1e31-45bb-a3ce-aef0402f6489 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.818598] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e48b21-a7e0-4ea7-9f62-4274dd0a66e2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.853216] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39694f71-b046-4371-aa01-eccc207e4503 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.866069] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62658474-cafa-4d6a-ba62-2db57cd9570b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.883777] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5223e974-e608-de00-a80d-fb21c10467e8, 'name': SearchDatastore_Task, 'duration_secs': 0.065856} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.884262] env[65121]: DEBUG nova.compute.provider_tree [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1244.887416] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a301db59-2a5d-48d0-8d8f-243f5ab8d2e4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.893396] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1244.893396] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52915c11-4c7a-e365-d4b2-c00d79ad80b1" [ 1244.893396] env[65121]: _type = "Task" [ 1244.893396] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.906299] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52915c11-4c7a-e365-d4b2-c00d79ad80b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.188191] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1245.188591] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ddface42-f751-4c12-9f29-5725b378eb7b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.198626] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1245.198626] env[65121]: value = "task-5107570" [ 1245.198626] env[65121]: _type = "Task" [ 1245.198626] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.209909] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.267946] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107569, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.389701] env[65121]: DEBUG nova.scheduler.client.report [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1245.410503] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52915c11-4c7a-e365-d4b2-c00d79ad80b1, 'name': SearchDatastore_Task, 'duration_secs': 0.024085} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.411791] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1245.412197] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b669fb5d-a68f-4390-b903-80879c239b4e/b669fb5d-a68f-4390-b903-80879c239b4e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1245.412608] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23801473-c015-4f94-b8ea-02cbb45b3a3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.423461] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1245.423461] env[65121]: value = "task-5107571" [ 1245.423461] env[65121]: _type = "Task" [ 1245.423461] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.438644] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.461437] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.461713] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1245.462030] env[65121]: INFO nova.compute.manager [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Attaching volume 0fa3fc8e-7557-4c45-96de-ab66ddf12df4 to /dev/sdc [ 1245.503652] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87080743-1138-4fed-8fd5-eb74b65bfee3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.517529] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab9233f-6b6e-4c93-a7cc-2362c71d32c9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.535836] env[65121]: DEBUG nova.virt.block_device [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updating existing volume attachment record: 9123551d-10b9-43b2-861b-fb55b5d7a1ba {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1245.712866] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.769709] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107569, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.901428] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1245.932801] env[65121]: INFO nova.scheduler.client.report [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted allocations for instance 138e48fe-0221-4173-801b-ee53b2bb98df [ 1245.946551] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.214996] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107570, 'name': PowerOffVM_Task, 'duration_secs': 0.930292} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.215383] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1246.215468] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance '91a36110-5e97-4553-90c3-3e4213e30c14' progress to 17 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1246.269878] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107569, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.437278] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.443482] env[65121]: DEBUG oslo_concurrency.lockutils [None req-420ff303-ff7b-4ae0-a5a9-34cd585af1e1 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "138e48fe-0221-4173-801b-ee53b2bb98df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.041s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1246.723437] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1246.723655] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1246.723881] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1246.724142] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1246.724302] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1246.725036] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1246.725036] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1246.725036] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1246.725276] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1246.725460] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1246.725679] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1246.732375] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8e89fe3-b357-472b-ab22-bfd51b994687 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.752086] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1246.752086] env[65121]: value = "task-5107573" [ 1246.752086] env[65121]: _type = "Task" [ 1246.752086] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.765386] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107573, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.771794] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107569, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.936549] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.266027] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107573, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.275714] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107569, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.425212} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.276117] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ceda632d-08b0-45f0-ae23-8456ddb4144e/ceda632d-08b0-45f0-ae23-8456ddb4144e.vmdk to [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1247.277056] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459fb910-9ac7-4b15-bbfc-b72425e06478 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.304275] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1247.304725] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9650c93-3784-4756-8361-f22b3d6403cc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.335163] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1247.335163] env[65121]: value = "task-5107574" [ 1247.335163] env[65121]: _type = "Task" [ 1247.335163] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.344958] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.440269] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107571, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.764188] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107573, 'name': ReconfigVM_Task, 'duration_secs': 0.572348} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.764520] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance '91a36110-5e97-4553-90c3-3e4213e30c14' progress to 33 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1247.844360] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107574, 'name': ReconfigVM_Task, 'duration_secs': 0.368519} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.844629] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3/999ab257-c116-4a73-9b87-b52aeb9a3bb3.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.845263] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0a9899b-8c37-47ec-b250-9a45177de139 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.852275] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1247.852275] env[65121]: value = "task-5107575" [ 1247.852275] env[65121]: _type = "Task" [ 1247.852275] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.860411] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107575, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.937691] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107571, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.097807} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.939885] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] b669fb5d-a68f-4390-b903-80879c239b4e/b669fb5d-a68f-4390-b903-80879c239b4e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1247.939885] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1247.939885] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a37f1a7-cd08-4c89-8cc5-955639636ab0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.945844] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1247.945844] env[65121]: value = "task-5107576" [ 1247.945844] env[65121]: _type = "Task" [ 1247.945844] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.955342] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107576, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.271868] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1248.271868] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1248.271868] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1248.272400] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1248.272400] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1248.272474] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1248.273315] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1248.273315] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1248.273315] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1248.273315] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1248.273525] env[65121]: DEBUG nova.virt.hardware [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1248.279072] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfiguring VM instance instance-0000007b to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1248.279381] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-191e4714-ca80-49fa-bd8b-2fbe4e8bd4d3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.299517] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1248.299517] env[65121]: value = "task-5107577" [ 1248.299517] env[65121]: _type = "Task" [ 1248.299517] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.308564] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107577, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.363136] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107575, 'name': Rename_Task, 'duration_secs': 0.163907} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.363419] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1248.363738] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25af0d9c-724b-4e2e-bb24-c51ba293a387 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.370963] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1248.370963] env[65121]: value = "task-5107578" [ 1248.370963] env[65121]: _type = "Task" [ 1248.370963] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.379124] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.457960] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07044} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.458268] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1248.459121] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1273b64a-22c1-4d96-b606-1305d7bcaa09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.483479] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Reconfiguring VM instance instance-0000007c to attach disk [datastore1] b669fb5d-a68f-4390-b903-80879c239b4e/b669fb5d-a68f-4390-b903-80879c239b4e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1248.483789] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95f8600d-3fba-4850-b8cc-381d6782fdb0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.504503] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1248.504503] env[65121]: value = "task-5107579" [ 1248.504503] env[65121]: _type = "Task" [ 1248.504503] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.513282] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107579, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.807093] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.807340] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.815851] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107577, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.881017] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107578, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.017911] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107579, 'name': ReconfigVM_Task, 'duration_secs': 0.278649} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.018160] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Reconfigured VM instance instance-0000007c to attach disk [datastore1] b669fb5d-a68f-4390-b903-80879c239b4e/b669fb5d-a68f-4390-b903-80879c239b4e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1249.018795] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49bf2a67-2394-47f0-96f5-22710ba80363 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.025811] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1249.025811] env[65121]: value = "task-5107580" [ 1249.025811] env[65121]: _type = "Task" [ 1249.025811] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.034659] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107580, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.310508] env[65121]: DEBUG nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1249.313203] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107577, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.381491] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107578, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.536157] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107580, 'name': Rename_Task, 'duration_secs': 0.15569} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.536432] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1249.536685] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71118c32-095f-43dd-a306-98867cdf0e37 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.544493] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1249.544493] env[65121]: value = "task-5107581" [ 1249.544493] env[65121]: _type = "Task" [ 1249.544493] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.553114] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107581, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.811592] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107577, 'name': ReconfigVM_Task, 'duration_secs': 1.177408} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.811860] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfigured VM instance instance-0000007b to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1249.812694] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739a2bb8-33a7-46cd-a132-92ebdb5002a2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.841390] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] volume-47c28970-ebc3-443e-a72c-4ac58b364e40/volume-47c28970-ebc3-443e-a72c-4ac58b364e40.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1249.842811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1249.843086] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1249.844684] env[65121]: INFO nova.compute.claims [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1249.848510] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48fd8053-1e54-4f95-af74-73c66308386a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.868332] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1249.868332] env[65121]: value = "task-5107583" [ 1249.868332] env[65121]: _type = "Task" [ 1249.868332] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.878197] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107583, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.883705] env[65121]: DEBUG oslo_vmware.api [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107578, 'name': PowerOnVM_Task, 'duration_secs': 1.244017} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.884258] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1250.020305] env[65121]: DEBUG nova.compute.manager [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1250.021834] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd85b3f9-6b5a-432d-bf8f-c24bc5321ced {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.055818] env[65121]: DEBUG oslo_vmware.api [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107581, 'name': PowerOnVM_Task, 'duration_secs': 0.508882} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.056112] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1250.056317] env[65121]: INFO nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Took 9.22 seconds to spawn the instance on the hypervisor. [ 1250.056489] env[65121]: DEBUG nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1250.057296] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55127a87-87f2-40c6-856d-f82ba68ebe82 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.378990] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107583, 'name': ReconfigVM_Task, 'duration_secs': 0.281347} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.379401] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfigured VM instance instance-0000007b to attach disk [datastore1] volume-47c28970-ebc3-443e-a72c-4ac58b364e40/volume-47c28970-ebc3-443e-a72c-4ac58b364e40.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1250.379529] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance '91a36110-5e97-4553-90c3-3e4213e30c14' progress to 50 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1250.543890] env[65121]: DEBUG oslo_concurrency.lockutils [None req-49391574-e3b6-4f16-9346-a5f218832c3e tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.477s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1250.575895] env[65121]: INFO nova.compute.manager [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Took 14.09 seconds to build instance. [ 1250.886227] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7100a9ff-cca0-49a7-a839-c591063c4c1f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.912354] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c7226f-de3d-4d33-a155-ff59267fb475 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.930980] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance '91a36110-5e97-4553-90c3-3e4213e30c14' progress to 67 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1251.065435] env[65121]: DEBUG nova.compute.manager [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Received event network-changed-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1251.065435] env[65121]: DEBUG nova.compute.manager [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Refreshing instance network info cache due to event network-changed-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1251.065627] env[65121]: DEBUG oslo_concurrency.lockutils [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Acquiring lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.065741] env[65121]: DEBUG oslo_concurrency.lockutils [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Acquired lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1251.065928] env[65121]: DEBUG nova.network.neutron [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Refreshing network info cache for port da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1251.076938] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d53e4eb9-a56f-46fc-ae29-8e81dd0e4a9d tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.599s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1251.082459] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bba5a3-0156-4dd4-a1c8-08d3c01cf75c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.092368] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45d9ab4-b342-41ab-ae48-be21c46f20e5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.125257] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71afa9ac-d027-40aa-bd73-44b05cc37649 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.134567] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9231b71-ff6c-4967-988f-7678d768765b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.150130] env[65121]: DEBUG nova.compute.provider_tree [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.572611] env[65121]: WARNING neutronclient.v2_0.client [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1251.573458] env[65121]: WARNING openstack [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.573881] env[65121]: WARNING openstack [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.653857] env[65121]: DEBUG nova.scheduler.client.report [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1251.696800] env[65121]: WARNING openstack [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.697191] env[65121]: WARNING openstack [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.763065] env[65121]: WARNING neutronclient.v2_0.client [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1251.763738] env[65121]: WARNING openstack [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.764119] env[65121]: WARNING openstack [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.848076] env[65121]: DEBUG nova.network.neutron [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updated VIF entry in instance network info cache for port da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1251.848462] env[65121]: DEBUG nova.network.neutron [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updating instance_info_cache with network_info: [{"id": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "address": "fa:16:3e:2b:98:34", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d86d6-62", "ovs_interfaceid": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1252.090457] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1252.090715] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993601', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'name': 'volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'serial': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1252.091661] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51529ffc-5407-4142-a9bd-9d5bf644086d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.110412] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5747fe0d-f6d2-4e71-8820-a2b3f75eceb7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.138410] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4/volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1252.138722] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8bd1690-1fa4-4157-b65f-dc6b82fe739d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.158674] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1252.159203] env[65121]: DEBUG nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1252.163511] env[65121]: DEBUG oslo_vmware.api [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1252.163511] env[65121]: value = "task-5107584" [ 1252.163511] env[65121]: _type = "Task" [ 1252.163511] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.175932] env[65121]: DEBUG oslo_vmware.api [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107584, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.351219] env[65121]: DEBUG oslo_concurrency.lockutils [req-d987a7c2-ac1f-4fa2-90f5-7c16e3c9c49a req-237b810f-b8ed-4a05-bccb-4fa792a86c33 service nova] Releasing lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1252.559485] env[65121]: WARNING neutronclient.v2_0.client [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1252.618372] env[65121]: DEBUG nova.network.neutron [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Port 76e75737-0e97-46ba-89f7-b94bbcbff126 binding to destination host cpu-1 is already ACTIVE {{(pid=65121) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3270}} [ 1252.663940] env[65121]: DEBUG nova.compute.utils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1252.665551] env[65121]: DEBUG nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1252.669124] env[65121]: DEBUG nova.network.neutron [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1252.669530] env[65121]: WARNING neutronclient.v2_0.client [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1252.669895] env[65121]: WARNING neutronclient.v2_0.client [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1252.670504] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1252.670849] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1252.695027] env[65121]: DEBUG oslo_vmware.api [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107584, 'name': ReconfigVM_Task, 'duration_secs': 0.374169} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.695316] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfigured VM instance instance-00000075 to attach disk [datastore1] volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4/volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1252.701130] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2be9532-4c01-4006-9733-b9b9cf145c9a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.719489] env[65121]: DEBUG oslo_vmware.api [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1252.719489] env[65121]: value = "task-5107585" [ 1252.719489] env[65121]: _type = "Task" [ 1252.719489] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.729223] env[65121]: DEBUG oslo_vmware.api [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107585, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.736380] env[65121]: DEBUG nova.policy [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4853c778ab79419da0e9de23ee3617a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f056059180af48bba4587006efb0b211', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1253.072258] env[65121]: DEBUG nova.network.neutron [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Successfully created port: 1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1253.182668] env[65121]: DEBUG nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1253.230273] env[65121]: DEBUG oslo_vmware.api [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107585, 'name': ReconfigVM_Task, 'duration_secs': 0.157069} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.230602] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993601', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'name': 'volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'serial': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1253.640599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1253.640599] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1253.641084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1254.194389] env[65121]: DEBUG nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1254.222939] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1254.223198] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1254.223346] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1254.223515] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1254.223653] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1254.223792] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1254.224024] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1254.224236] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1254.224407] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1254.224563] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1254.224825] env[65121]: DEBUG nova.virt.hardware [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1254.225744] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aebdf22-a20a-4ddc-aa2e-90bdc9a7db2a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.234802] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0867bfa-c96e-4451-bd3d-5642f10a3c99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.269759] env[65121]: DEBUG nova.objects.instance [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1254.491716] env[65121]: DEBUG nova.compute.manager [req-c23a404d-69bb-41db-a543-23f33b94305e req-a9d94dd8-6c24-4938-9b60-3878b1ecd949 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Received event network-vif-plugged-1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1254.492028] env[65121]: DEBUG oslo_concurrency.lockutils [req-c23a404d-69bb-41db-a543-23f33b94305e req-a9d94dd8-6c24-4938-9b60-3878b1ecd949 service nova] Acquiring lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1254.492314] env[65121]: DEBUG oslo_concurrency.lockutils [req-c23a404d-69bb-41db-a543-23f33b94305e req-a9d94dd8-6c24-4938-9b60-3878b1ecd949 service nova] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1254.492544] env[65121]: DEBUG oslo_concurrency.lockutils [req-c23a404d-69bb-41db-a543-23f33b94305e req-a9d94dd8-6c24-4938-9b60-3878b1ecd949 service nova] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1254.492778] env[65121]: DEBUG nova.compute.manager [req-c23a404d-69bb-41db-a543-23f33b94305e req-a9d94dd8-6c24-4938-9b60-3878b1ecd949 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] No waiting events found dispatching network-vif-plugged-1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1254.493169] env[65121]: WARNING nova.compute.manager [req-c23a404d-69bb-41db-a543-23f33b94305e req-a9d94dd8-6c24-4938-9b60-3878b1ecd949 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Received unexpected event network-vif-plugged-1603a551-b060-4f2e-a724-cfcd68db842a for instance with vm_state building and task_state spawning. [ 1254.580865] env[65121]: DEBUG nova.network.neutron [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Successfully updated port: 1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1254.647370] env[65121]: WARNING neutronclient.v2_0.client [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1254.679347] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.679524] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1254.679720] env[65121]: DEBUG nova.network.neutron [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1254.774676] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d67f9338-57eb-4584-845d-2708621c531f tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.313s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1255.084348] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.084348] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1255.084348] env[65121]: DEBUG nova.network.neutron [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1255.172044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.172044] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.182676] env[65121]: WARNING neutronclient.v2_0.client [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1255.183710] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1255.183851] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1255.299821] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1255.300386] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1255.360061] env[65121]: WARNING neutronclient.v2_0.client [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1255.360740] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1255.361108] env[65121]: WARNING openstack [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1255.446759] env[65121]: DEBUG nova.network.neutron [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [{"id": "76e75737-0e97-46ba-89f7-b94bbcbff126", "address": "fa:16:3e:38:42:7e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e75737-0e", "ovs_interfaceid": "76e75737-0e97-46ba-89f7-b94bbcbff126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1255.588042] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1255.588042] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1255.623463] env[65121]: DEBUG nova.network.neutron [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1255.643361] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1255.643730] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1255.676179] env[65121]: INFO nova.compute.manager [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Detaching volume 5d4df019-195e-4afc-bd9f-7e8d84c011f2 [ 1255.720871] env[65121]: WARNING neutronclient.v2_0.client [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1255.720871] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1255.720871] env[65121]: WARNING openstack [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1255.748537] env[65121]: INFO nova.virt.block_device [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Attempting to driver detach volume 5d4df019-195e-4afc-bd9f-7e8d84c011f2 from mountpoint /dev/sdb [ 1255.748537] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1255.748537] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993597', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'name': 'volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'serial': '5d4df019-195e-4afc-bd9f-7e8d84c011f2'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1255.749490] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21f2a8e-2bb9-415b-a011-d6331c4bb6e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.775083] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62e53ae-002f-4aee-bfcc-36e36ac4d8c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.786416] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add1ac99-2106-47f8-a27c-6386535be348 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.811839] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf440fe3-d7b3-4a72-bc36-231824117635 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.829112] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The volume has not been displaced from its original location: [datastore1] volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2/volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1255.835148] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfiguring VM instance instance-00000075 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1255.838305] env[65121]: DEBUG nova.network.neutron [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updating instance_info_cache with network_info: [{"id": "1603a551-b060-4f2e-a724-cfcd68db842a", "address": "fa:16:3e:2c:39:10", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1603a551-b0", "ovs_interfaceid": "1603a551-b060-4f2e-a724-cfcd68db842a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1255.840024] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d65aa2ab-0240-4ebc-82c6-84637ea37fd7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.860519] env[65121]: DEBUG oslo_vmware.api [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1255.860519] env[65121]: value = "task-5107586" [ 1255.860519] env[65121]: _type = "Task" [ 1255.860519] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.874019] env[65121]: DEBUG oslo_vmware.api [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107586, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.924829] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.925129] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.925306] env[65121]: INFO nova.compute.manager [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Shelving [ 1255.949672] env[65121]: DEBUG oslo_concurrency.lockutils [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1256.341398] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1256.342225] env[65121]: DEBUG nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Instance network_info: |[{"id": "1603a551-b060-4f2e-a724-cfcd68db842a", "address": "fa:16:3e:2c:39:10", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1603a551-b0", "ovs_interfaceid": "1603a551-b060-4f2e-a724-cfcd68db842a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1256.342405] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:39:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1603a551-b060-4f2e-a724-cfcd68db842a', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1256.351206] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1256.351495] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1256.351805] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6dcd931f-6305-455f-af45-740803fa2131 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.376333] env[65121]: DEBUG oslo_vmware.api [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107586, 'name': ReconfigVM_Task, 'duration_secs': 0.237049} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.377584] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfigured VM instance instance-00000075 to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1256.382195] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1256.382195] env[65121]: value = "task-5107587" [ 1256.382195] env[65121]: _type = "Task" [ 1256.382195] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.382401] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfa7b7f8-e172-4c71-8d70-a44caa4acfd1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.401492] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107587, 'name': CreateVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.402947] env[65121]: DEBUG oslo_vmware.api [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1256.402947] env[65121]: value = "task-5107588" [ 1256.402947] env[65121]: _type = "Task" [ 1256.402947] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.411571] env[65121]: DEBUG oslo_vmware.api [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.460256] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f517f28f-284d-4418-9d9f-de6ff699da82 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.468301] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec75d3c-a6a1-468c-92a2-0941f21e825f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.515596] env[65121]: DEBUG nova.compute.manager [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Received event network-changed-1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1256.515803] env[65121]: DEBUG nova.compute.manager [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Refreshing instance network info cache due to event network-changed-1603a551-b060-4f2e-a724-cfcd68db842a. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1256.516071] env[65121]: DEBUG oslo_concurrency.lockutils [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Acquiring lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.516217] env[65121]: DEBUG oslo_concurrency.lockutils [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Acquired lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1256.516415] env[65121]: DEBUG nova.network.neutron [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Refreshing network info cache for port 1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1256.905039] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107587, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.914808] env[65121]: DEBUG oslo_vmware.api [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107588, 'name': ReconfigVM_Task, 'duration_secs': 0.163636} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.915132] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993597', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'name': 'volume-5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '5d4df019-195e-4afc-bd9f-7e8d84c011f2', 'serial': '5d4df019-195e-4afc-bd9f-7e8d84c011f2'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1256.935368] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1256.935687] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fdc0ba1-9ae2-4e3f-9d0a-6234c95c94c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.942699] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1256.942699] env[65121]: value = "task-5107589" [ 1256.942699] env[65121]: _type = "Task" [ 1256.942699] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.951470] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.019462] env[65121]: WARNING neutronclient.v2_0.client [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1257.020337] env[65121]: WARNING openstack [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1257.020873] env[65121]: WARNING openstack [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.131556] env[65121]: WARNING openstack [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1257.131998] env[65121]: WARNING openstack [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.197907] env[65121]: WARNING neutronclient.v2_0.client [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1257.198652] env[65121]: WARNING openstack [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1257.199057] env[65121]: WARNING openstack [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.300784] env[65121]: DEBUG nova.network.neutron [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updated VIF entry in instance network info cache for port 1603a551-b060-4f2e-a724-cfcd68db842a. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1257.301147] env[65121]: DEBUG nova.network.neutron [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updating instance_info_cache with network_info: [{"id": "1603a551-b060-4f2e-a724-cfcd68db842a", "address": "fa:16:3e:2c:39:10", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1603a551-b0", "ovs_interfaceid": "1603a551-b060-4f2e-a724-cfcd68db842a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1257.404734] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107587, 'name': CreateVM_Task, 'duration_secs': 0.5568} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.404960] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1257.405439] env[65121]: WARNING neutronclient.v2_0.client [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1257.405791] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.405939] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1257.406280] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1257.406570] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71aaeb2c-de43-43bd-8cee-2706bb09e8dc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.411999] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1257.411999] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522f38c0-5865-4aea-f9d7-8971b3531746" [ 1257.411999] env[65121]: _type = "Task" [ 1257.411999] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.420394] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522f38c0-5865-4aea-f9d7-8971b3531746, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.453240] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107589, 'name': PowerOffVM_Task, 'duration_secs': 0.260541} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.453360] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1257.454138] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e27228-744e-4072-9ec9-d81f45c75c02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.475398] env[65121]: DEBUG nova.objects.instance [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1257.478110] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f967dc-8144-499c-929d-7fee08ca64da {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.573272] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5343ba8-3f53-4689-9bed-bfd8658712fa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.592909] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb02cad-b16e-4c3d-9664-70c009ae15a6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.600193] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance '91a36110-5e97-4553-90c3-3e4213e30c14' progress to 83 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1257.804225] env[65121]: DEBUG oslo_concurrency.lockutils [req-da4ac30d-f751-4dcb-8aab-0f83b93920c8 req-84cf2ff0-8c3e-49a7-ad98-e444caf45b04 service nova] Releasing lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1257.923049] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522f38c0-5865-4aea-f9d7-8971b3531746, 'name': SearchDatastore_Task, 'duration_secs': 0.010666} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.923049] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1257.923397] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1257.923505] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.923646] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1257.923821] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1257.924106] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47a05e83-df75-4af8-bb3f-5bcda6cb7258 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.936231] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1257.936432] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1257.937229] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb83666d-85ff-4a05-9c74-831913fbc286 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.943346] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1257.943346] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52873818-581a-ce98-dfb7-967591362354" [ 1257.943346] env[65121]: _type = "Task" [ 1257.943346] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.956216] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52873818-581a-ce98-dfb7-967591362354, 'name': SearchDatastore_Task, 'duration_secs': 0.010643} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.960068] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64681836-ecf7-4f7e-ac3c-493654d1b31c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.967520] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1257.967520] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5246bf56-8c4c-78df-6756-2c1f6b15415b" [ 1257.967520] env[65121]: _type = "Task" [ 1257.967520] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.977646] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246bf56-8c4c-78df-6756-2c1f6b15415b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.990280] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Creating Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1257.990812] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-673e1f0f-67e1-4bac-9fc4-80cea9f7f996 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.999804] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1257.999804] env[65121]: value = "task-5107590" [ 1257.999804] env[65121]: _type = "Task" [ 1257.999804] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.008661] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107590, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.105858] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1258.106186] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2a06ccb-de77-42fe-beae-10d5d4ccf5a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.113844] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1258.113844] env[65121]: value = "task-5107591" [ 1258.113844] env[65121]: _type = "Task" [ 1258.113844] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.122667] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107591, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.464163] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1258.479061] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5246bf56-8c4c-78df-6756-2c1f6b15415b, 'name': SearchDatastore_Task, 'duration_secs': 0.015862} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.479386] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1258.479697] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 07ff0535-f83c-4a1b-8077-26a8b1fa02b2/07ff0535-f83c-4a1b-8077-26a8b1fa02b2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1258.479984] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e73734e7-5bfa-41dc-95e1-ecd45f268ddb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.482561] env[65121]: DEBUG oslo_concurrency.lockutils [None req-8868b7b3-2750-44e9-9dcc-463792872d1d tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.311s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1258.483855] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.020s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1258.492561] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1258.492561] env[65121]: value = "task-5107592" [ 1258.492561] env[65121]: _type = "Task" [ 1258.492561] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.505477] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107592, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.515940] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107590, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.625074] env[65121]: DEBUG oslo_vmware.api [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107591, 'name': PowerOnVM_Task, 'duration_secs': 0.469593} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.625074] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1258.625339] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-abe5ba6f-16fc-4b46-bb17-0430724b73b6 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance '91a36110-5e97-4553-90c3-3e4213e30c14' progress to 100 {{(pid=65121) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1258.989241] env[65121]: INFO nova.compute.manager [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Detaching volume 0fa3fc8e-7557-4c45-96de-ab66ddf12df4 [ 1259.006741] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107592, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.016959] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107590, 'name': CreateSnapshot_Task, 'duration_secs': 0.730824} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.017290] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Created Snapshot of the VM instance {{(pid=65121) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1259.018088] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79503b94-e8c7-43e0-ad4d-869bd92403d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.032194] env[65121]: INFO nova.virt.block_device [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Attempting to driver detach volume 0fa3fc8e-7557-4c45-96de-ab66ddf12df4 from mountpoint /dev/sdc [ 1259.032194] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1259.032194] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993601', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'name': 'volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'serial': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1259.032938] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a94c74-6232-45e3-be85-bc51801c0665 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.055729] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f80f50-5a74-45ba-91d4-9567df70d53d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.064072] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658e431c-1238-490b-9409-412ba24bd94a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.085269] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f27a207-d258-49ab-8429-f740ca417a0a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.100896] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] The volume has not been displaced from its original location: [datastore1] volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4/volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1259.106071] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfiguring VM instance instance-00000075 to detach disk 2002 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1259.106355] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb83a580-e594-4d07-b4ef-a94e3e22b00f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.124609] env[65121]: DEBUG oslo_vmware.api [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1259.124609] env[65121]: value = "task-5107593" [ 1259.124609] env[65121]: _type = "Task" [ 1259.124609] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.135919] env[65121]: DEBUG oslo_vmware.api [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107593, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.505357] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107592, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578826} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.505640] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 07ff0535-f83c-4a1b-8077-26a8b1fa02b2/07ff0535-f83c-4a1b-8077-26a8b1fa02b2.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1259.505854] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1259.506242] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3097dd5e-58ea-4ba4-b943-b1427151579a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.514789] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1259.514789] env[65121]: value = "task-5107594" [ 1259.514789] env[65121]: _type = "Task" [ 1259.514789] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.523485] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107594, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.543786] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Creating linked-clone VM from snapshot {{(pid=65121) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1259.544179] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-aebd5cf0-f6f6-47ee-9129-910dbc8c796d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.553251] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1259.553251] env[65121]: value = "task-5107595" [ 1259.553251] env[65121]: _type = "Task" [ 1259.553251] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.563493] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107595, 'name': CloneVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.635298] env[65121]: DEBUG oslo_vmware.api [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107593, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.024964] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107594, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096336} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.025356] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1260.026034] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f297e590-c7da-4533-bef8-aa85bc5ace93 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.049160] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Reconfiguring VM instance instance-0000007d to attach disk [datastore1] 07ff0535-f83c-4a1b-8077-26a8b1fa02b2/07ff0535-f83c-4a1b-8077-26a8b1fa02b2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.049523] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebfb7e3b-e0f2-48cd-91e3-8de402a4689c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.073686] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107595, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.075157] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1260.075157] env[65121]: value = "task-5107596" [ 1260.075157] env[65121]: _type = "Task" [ 1260.075157] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.083664] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107596, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.137567] env[65121]: DEBUG oslo_vmware.api [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107593, 'name': ReconfigVM_Task, 'duration_secs': 0.769036} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.137980] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Reconfigured VM instance instance-00000075 to detach disk 2002 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1260.145844] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f671a2d-5793-4baf-9790-2a3bb5e76a62 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.171438] env[65121]: DEBUG oslo_vmware.api [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1260.171438] env[65121]: value = "task-5107597" [ 1260.171438] env[65121]: _type = "Task" [ 1260.171438] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.186441] env[65121]: DEBUG oslo_vmware.api [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107597, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.576521] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107595, 'name': CloneVM_Task} progress is 94%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.587122] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107596, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.685085] env[65121]: DEBUG oslo_vmware.api [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107597, 'name': ReconfigVM_Task, 'duration_secs': 0.172597} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.685560] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993601', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'name': 'volume-0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2', 'attached_at': '', 'detached_at': '', 'volume_id': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4', 'serial': '0fa3fc8e-7557-4c45-96de-ab66ddf12df4'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1261.075744] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107595, 'name': CloneVM_Task, 'duration_secs': 1.395618} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.075999] env[65121]: INFO nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Created linked-clone VM from snapshot [ 1261.077057] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac29c6d-00f5-4a98-80e4-f931bba2087a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.094510] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Uploading image b118bbe9-8659-4100-a942-549907f6e6cf {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1261.100451] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107596, 'name': ReconfigVM_Task, 'duration_secs': 0.565301} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.101287] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Reconfigured VM instance instance-0000007d to attach disk [datastore1] 07ff0535-f83c-4a1b-8077-26a8b1fa02b2/07ff0535-f83c-4a1b-8077-26a8b1fa02b2.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.102029] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61025ff8-4950-40fe-83d1-f313eff2b23e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.110508] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1261.110508] env[65121]: value = "task-5107598" [ 1261.110508] env[65121]: _type = "Task" [ 1261.110508] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.120154] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107598, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.126140] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1261.126140] env[65121]: value = "vm-993604" [ 1261.126140] env[65121]: _type = "VirtualMachine" [ 1261.126140] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1261.126436] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4693ac1a-c017-4932-afa7-bcb60877620a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.135180] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease: (returnval){ [ 1261.135180] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab6ffd-6d33-ae27-50bb-c337d11e53a8" [ 1261.135180] env[65121]: _type = "HttpNfcLease" [ 1261.135180] env[65121]: } obtained for exporting VM: (result){ [ 1261.135180] env[65121]: value = "vm-993604" [ 1261.135180] env[65121]: _type = "VirtualMachine" [ 1261.135180] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1261.135577] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the lease: (returnval){ [ 1261.135577] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab6ffd-6d33-ae27-50bb-c337d11e53a8" [ 1261.135577] env[65121]: _type = "HttpNfcLease" [ 1261.135577] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1261.143398] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1261.143398] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab6ffd-6d33-ae27-50bb-c337d11e53a8" [ 1261.143398] env[65121]: _type = "HttpNfcLease" [ 1261.143398] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1261.234046] env[65121]: DEBUG nova.objects.instance [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'flavor' on Instance uuid b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1261.452541] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "91a36110-5e97-4553-90c3-3e4213e30c14" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1261.452541] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1261.452541] env[65121]: DEBUG nova.compute.manager [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Going to confirm migration 10 {{(pid=65121) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1261.597852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "e19d6d9b-d15d-4780-8564-f479573ce59b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1261.598143] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1261.598355] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "e19d6d9b-d15d-4780-8564-f479573ce59b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1261.598527] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1261.598688] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.600946] env[65121]: INFO nova.compute.manager [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Terminating instance [ 1261.621451] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107598, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.643194] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1261.643194] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab6ffd-6d33-ae27-50bb-c337d11e53a8" [ 1261.643194] env[65121]: _type = "HttpNfcLease" [ 1261.643194] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1261.643509] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1261.643509] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ab6ffd-6d33-ae27-50bb-c337d11e53a8" [ 1261.643509] env[65121]: _type = "HttpNfcLease" [ 1261.643509] env[65121]: }. {{(pid=65121) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1261.644219] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11440940-6269-48bd-84c6-25176a7a800c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.652155] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52824814-eaa7-c888-05ee-16bd55b3b44c/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1261.652311] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52824814-eaa7-c888-05ee-16bd55b3b44c/disk-0.vmdk for reading. {{(pid=65121) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1261.813376] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-74bf3198-6dbf-427c-bc55-9b1505e5338e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.949167] env[65121]: WARNING neutronclient.v2_0.client [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1261.992325] env[65121]: WARNING neutronclient.v2_0.client [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1261.992669] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1261.992813] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquired lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1261.992977] env[65121]: DEBUG nova.network.neutron [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1261.993169] env[65121]: DEBUG nova.objects.instance [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'info_cache' on Instance uuid 91a36110-5e97-4553-90c3-3e4213e30c14 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1262.105180] env[65121]: DEBUG nova.compute.manager [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1262.105601] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1262.106654] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7427a0-e242-44d3-9667-9af43c31ef70 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.118598] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1262.119390] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-557e051d-0801-4a82-926d-dd1a3f061a38 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.124190] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107598, 'name': Rename_Task, 'duration_secs': 0.760494} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.124440] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1262.124706] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ec2f78b-b86f-4e9b-b385-02213998fbd3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.127282] env[65121]: DEBUG oslo_vmware.api [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1262.127282] env[65121]: value = "task-5107600" [ 1262.127282] env[65121]: _type = "Task" [ 1262.127282] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.135019] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1262.135019] env[65121]: value = "task-5107601" [ 1262.135019] env[65121]: _type = "Task" [ 1262.135019] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.141693] env[65121]: DEBUG oslo_vmware.api [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.148017] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107601, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.241915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d133d062-8399-49f6-92c7-e4f662d54533 tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.758s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1263.152451] env[65121]: DEBUG oslo_vmware.api [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107601, 'name': PowerOnVM_Task, 'duration_secs': 0.532099} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.152955] env[65121]: DEBUG oslo_vmware.api [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107600, 'name': PowerOffVM_Task, 'duration_secs': 0.197262} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.153101] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1263.153346] env[65121]: INFO nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Took 8.96 seconds to spawn the instance on the hypervisor. [ 1263.153544] env[65121]: DEBUG nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1263.153910] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1263.154179] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1263.155237] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aad480a-ff0c-458a-9490-897d3cff0a95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.158298] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c6333023-f576-4568-928d-8fef13918a6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.242146] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1263.243434] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1263.243434] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Deleting the datastore file [datastore2] e19d6d9b-d15d-4780-8564-f479573ce59b {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1263.243434] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49768502-5a73-4f14-b79c-759237c62821 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.253708] env[65121]: DEBUG oslo_vmware.api [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for the task: (returnval){ [ 1263.253708] env[65121]: value = "task-5107603" [ 1263.253708] env[65121]: _type = "Task" [ 1263.253708] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.262691] env[65121]: DEBUG oslo_vmware.api [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.634225] env[65121]: WARNING neutronclient.v2_0.client [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.635114] env[65121]: WARNING openstack [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.635820] env[65121]: WARNING openstack [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.643798] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.644152] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.644408] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.644645] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.644912] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1263.647126] env[65121]: INFO nova.compute.manager [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Terminating instance [ 1263.681395] env[65121]: INFO nova.compute.manager [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Took 13.86 seconds to build instance. [ 1263.766765] env[65121]: DEBUG oslo_vmware.api [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Task: {'id': task-5107603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.318824} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.767360] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1263.768644] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1263.768644] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1263.768644] env[65121]: INFO nova.compute.manager [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1263.768644] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1263.768644] env[65121]: DEBUG nova.compute.manager [-] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1263.768644] env[65121]: DEBUG nova.network.neutron [-] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1263.769466] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.770288] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.771910] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.796020] env[65121]: WARNING openstack [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.796715] env[65121]: WARNING openstack [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.825711] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.875695] env[65121]: WARNING neutronclient.v2_0.client [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.876376] env[65121]: WARNING openstack [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.876803] env[65121]: WARNING openstack [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1264.153538] env[65121]: DEBUG nova.compute.manager [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1264.154031] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1264.155183] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b359df0f-b009-4d68-af7c-0576bcffc02f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.165151] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1264.165828] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ace0f0e-5c2a-4032-8cd1-a4ca65f04288 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.174705] env[65121]: DEBUG oslo_vmware.api [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1264.174705] env[65121]: value = "task-5107604" [ 1264.174705] env[65121]: _type = "Task" [ 1264.174705] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.185680] env[65121]: DEBUG oslo_concurrency.lockutils [None req-090c0f58-99ef-430e-bbad-5b13a5346c09 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.378s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1264.186197] env[65121]: DEBUG oslo_vmware.api [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107604, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.306798] env[65121]: DEBUG nova.compute.manager [req-458b741f-12c9-4633-a2b4-741896c42845 req-be0d47c0-4394-450a-bcea-fbab566a9ff4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Received event network-vif-deleted-c6e3761c-d9e9-4e36-937f-2df79e305a63 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1264.307629] env[65121]: INFO nova.compute.manager [req-458b741f-12c9-4633-a2b4-741896c42845 req-be0d47c0-4394-450a-bcea-fbab566a9ff4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Neutron deleted interface c6e3761c-d9e9-4e36-937f-2df79e305a63; detaching it from the instance and deleting it from the info cache [ 1264.307629] env[65121]: DEBUG nova.network.neutron [req-458b741f-12c9-4633-a2b4-741896c42845 req-be0d47c0-4394-450a-bcea-fbab566a9ff4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1264.319272] env[65121]: DEBUG nova.network.neutron [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [{"id": "76e75737-0e97-46ba-89f7-b94bbcbff126", "address": "fa:16:3e:38:42:7e", "network": {"id": "dfa8cb96-73ef-4474-a41f-4c61c03338fe", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1896855897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00383aa3355e438cb703c2b86c7917f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e75737-0e", "ovs_interfaceid": "76e75737-0e97-46ba-89f7-b94bbcbff126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1264.513250] env[65121]: DEBUG nova.compute.manager [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Received event network-changed-1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1264.513383] env[65121]: DEBUG nova.compute.manager [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Refreshing instance network info cache due to event network-changed-1603a551-b060-4f2e-a724-cfcd68db842a. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1264.514541] env[65121]: DEBUG oslo_concurrency.lockutils [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Acquiring lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.514800] env[65121]: DEBUG oslo_concurrency.lockutils [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Acquired lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1264.514903] env[65121]: DEBUG nova.network.neutron [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Refreshing network info cache for port 1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1264.685294] env[65121]: DEBUG oslo_vmware.api [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107604, 'name': PowerOffVM_Task, 'duration_secs': 0.256991} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.685563] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1264.685719] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1264.685973] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a9c155a-7eda-405f-9322-a2b61f9c4a4c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.767794] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1264.768136] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1264.768392] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Deleting the datastore file [datastore1] b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1264.768715] env[65121]: DEBUG nova.network.neutron [-] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1264.770812] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d3bb46c-0f82-44c3-a651-cdf0dd8ec6f9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.779957] env[65121]: DEBUG oslo_vmware.api [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for the task: (returnval){ [ 1264.779957] env[65121]: value = "task-5107606" [ 1264.779957] env[65121]: _type = "Task" [ 1264.779957] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.790304] env[65121]: DEBUG oslo_vmware.api [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107606, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.811189] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bee628d4-965d-48a2-ac9c-e52b4acf32a5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.823811] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1847b9b-650d-40c4-855e-fb4f74b8e969 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.836039] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Releasing lock "refresh_cache-91a36110-5e97-4553-90c3-3e4213e30c14" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1264.836407] env[65121]: DEBUG nova.objects.instance [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'migration_context' on Instance uuid 91a36110-5e97-4553-90c3-3e4213e30c14 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1264.866652] env[65121]: DEBUG nova.compute.manager [req-458b741f-12c9-4633-a2b4-741896c42845 req-be0d47c0-4394-450a-bcea-fbab566a9ff4 service nova] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Detach interface failed, port_id=c6e3761c-d9e9-4e36-937f-2df79e305a63, reason: Instance e19d6d9b-d15d-4780-8564-f479573ce59b could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1265.017322] env[65121]: WARNING neutronclient.v2_0.client [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1265.020152] env[65121]: WARNING openstack [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.020152] env[65121]: WARNING openstack [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1265.137614] env[65121]: WARNING openstack [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.138158] env[65121]: WARNING openstack [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1265.211074] env[65121]: WARNING neutronclient.v2_0.client [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1265.211074] env[65121]: WARNING openstack [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.211074] env[65121]: WARNING openstack [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1265.273645] env[65121]: INFO nova.compute.manager [-] [instance: e19d6d9b-d15d-4780-8564-f479573ce59b] Took 1.50 seconds to deallocate network for instance. [ 1265.289970] env[65121]: DEBUG oslo_vmware.api [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Task: {'id': task-5107606, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361031} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.290257] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1265.290416] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1265.290579] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1265.290744] env[65121]: INFO nova.compute.manager [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1265.291041] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1265.291265] env[65121]: DEBUG nova.compute.manager [-] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1265.291367] env[65121]: DEBUG nova.network.neutron [-] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1265.291587] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1265.292180] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.292446] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1265.314820] env[65121]: DEBUG nova.network.neutron [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updated VIF entry in instance network info cache for port 1603a551-b060-4f2e-a724-cfcd68db842a. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1265.315226] env[65121]: DEBUG nova.network.neutron [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updating instance_info_cache with network_info: [{"id": "1603a551-b060-4f2e-a724-cfcd68db842a", "address": "fa:16:3e:2c:39:10", "network": {"id": "69f2d150-62cf-4c29-bfe9-c82e13e318a2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-345502435-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f056059180af48bba4587006efb0b211", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1603a551-b0", "ovs_interfaceid": "1603a551-b060-4f2e-a724-cfcd68db842a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1265.333928] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1265.341228] env[65121]: DEBUG nova.objects.base [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Object Instance<91a36110-5e97-4553-90c3-3e4213e30c14> lazy-loaded attributes: info_cache,migration_context {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1265.342581] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba8907c-774a-4bcf-a090-eaf1ddbb9905 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.365292] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbce398b-a77a-4b04-aa36-23edc8a2ad21 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.372955] env[65121]: DEBUG oslo_vmware.api [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1265.372955] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52a8ac90-2256-9b20-e586-1b82ed0248c6" [ 1265.372955] env[65121]: _type = "Task" [ 1265.372955] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.384035] env[65121]: DEBUG oslo_vmware.api [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a8ac90-2256-9b20-e586-1b82ed0248c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.785928] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1265.786272] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1265.786596] env[65121]: DEBUG nova.objects.instance [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lazy-loading 'resources' on Instance uuid e19d6d9b-d15d-4780-8564-f479573ce59b {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.818708] env[65121]: DEBUG oslo_concurrency.lockutils [req-11bfa748-c6d9-492e-9be5-649d1ca76dee req-22ba1802-a8c3-4ff8-984e-1640710e0a0f service nova] Releasing lock "refresh_cache-07ff0535-f83c-4a1b-8077-26a8b1fa02b2" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1265.884154] env[65121]: DEBUG oslo_vmware.api [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52a8ac90-2256-9b20-e586-1b82ed0248c6, 'name': SearchDatastore_Task, 'duration_secs': 0.009962} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.884578] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1266.343535] env[65121]: DEBUG nova.compute.manager [req-aca00635-f942-43cc-afdc-cf079d07429a req-d7fbda20-65d4-48a1-9376-342f89f94d16 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Received event network-vif-deleted-2dcc8fc2-ec38-4222-acf6-938b526306fa {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1266.343808] env[65121]: INFO nova.compute.manager [req-aca00635-f942-43cc-afdc-cf079d07429a req-d7fbda20-65d4-48a1-9376-342f89f94d16 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Neutron deleted interface 2dcc8fc2-ec38-4222-acf6-938b526306fa; detaching it from the instance and deleting it from the info cache [ 1266.344675] env[65121]: DEBUG nova.network.neutron [req-aca00635-f942-43cc-afdc-cf079d07429a req-d7fbda20-65d4-48a1-9376-342f89f94d16 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1266.490343] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ed6df0-32ee-4566-85f7-0c20af55c4c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.498745] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fa8842-9709-4c7c-bec2-d3935999ac45 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.531159] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2476665-fc5e-43cd-990c-666e5d13b51f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.539595] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789cdf4f-9351-4a50-93fb-3e567e51b716 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.555768] env[65121]: DEBUG nova.compute.provider_tree [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1266.727103] env[65121]: DEBUG nova.network.neutron [-] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1266.744535] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1266.744791] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1266.847968] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-876b127b-95f5-458b-9faf-6e74ea00d2fe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.861199] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ece8f6-cdea-46d6-8641-98a34f4eabff {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.897869] env[65121]: DEBUG nova.compute.manager [req-aca00635-f942-43cc-afdc-cf079d07429a req-d7fbda20-65d4-48a1-9376-342f89f94d16 service nova] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Detach interface failed, port_id=2dcc8fc2-ec38-4222-acf6-938b526306fa, reason: Instance b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1267.059838] env[65121]: DEBUG nova.scheduler.client.report [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1267.230568] env[65121]: INFO nova.compute.manager [-] [instance: b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2] Took 1.94 seconds to deallocate network for instance. [ 1267.248393] env[65121]: DEBUG nova.compute.utils [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1267.566345] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1267.568849] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.684s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1267.598367] env[65121]: INFO nova.scheduler.client.report [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Deleted allocations for instance e19d6d9b-d15d-4780-8564-f479573ce59b [ 1267.738209] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1267.751230] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1268.107628] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7921fbdb-5c52-4bca-b481-cb5f438dc431 tempest-ServersTestManualDisk-843760192 tempest-ServersTestManualDisk-843760192-project-member] Lock "e19d6d9b-d15d-4780-8564-f479573ce59b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.509s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1268.246052] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0d1826-563c-46aa-9f45-68cc49fd4327 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.254501] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1350749-562e-4e4e-a73d-a4f62757d4aa {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.288936] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14b7889-0c94-48fe-ae0f-039834ae4bd0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.298117] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af5bf14-c11c-41ad-8256-aab938815700 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.313507] env[65121]: DEBUG nova.compute.provider_tree [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1268.817265] env[65121]: DEBUG nova.scheduler.client.report [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1268.835210] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1268.835527] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1268.835767] env[65121]: INFO nova.compute.manager [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Attaching volume 50802f92-48fc-45a4-8fdb-0811fade10a3 to /dev/sdb [ 1268.876141] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c2c0b7-e662-4d7b-a8af-00c765128d8f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.887230] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2c80ec-2fa5-407c-9aa8-49d1b43ae4f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.903154] env[65121]: DEBUG nova.virt.block_device [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating existing volume attachment record: 1d337df4-abc3-4e4f-9f81-444d7e590281 {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1269.829052] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.260s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1269.833426] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.095s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.833810] env[65121]: DEBUG nova.objects.instance [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lazy-loading 'resources' on Instance uuid b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1270.276534] env[65121]: INFO nova.compute.manager [None req-c522aca0-390c-41b6-90e1-b784dafd2556 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Get console output [ 1270.276942] env[65121]: WARNING nova.virt.vmwareapi.driver [None req-c522aca0-390c-41b6-90e1-b784dafd2556 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] The console log is missing. Check your VSPC configuration [ 1270.420614] env[65121]: INFO nova.scheduler.client.report [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted allocation for migration 37985e1d-7fb7-436b-9d96-696e712f7a17 [ 1270.541580] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9db9d8-87f2-4e3b-b2c6-302db740e0e1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.550195] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b424d8-21bf-4e7f-8be9-4c2b1eec7bbc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.582509] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265a21b2-1aa2-4495-b4e8-ff7fe3703553 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.590528] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98da1ff-aad9-4308-a0e7-e35be3596c68 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.607112] env[65121]: DEBUG nova.compute.provider_tree [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.932137] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f0f7d4df-5937-42db-b51e-e8e4d990d7ed tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.490s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1270.940421] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52824814-eaa7-c888-05ee-16bd55b3b44c/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1270.941715] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b13ef4-df18-46f3-98bc-f1bf572ad0b1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.950466] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52824814-eaa7-c888-05ee-16bd55b3b44c/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1270.950705] env[65121]: ERROR oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52824814-eaa7-c888-05ee-16bd55b3b44c/disk-0.vmdk due to incomplete transfer. [ 1270.951761] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a5e36268-65f7-43e7-b569-747e289647a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.961850] env[65121]: DEBUG oslo_vmware.rw_handles [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52824814-eaa7-c888-05ee-16bd55b3b44c/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1270.962169] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Uploaded image b118bbe9-8659-4100-a942-549907f6e6cf to the Glance image server {{(pid=65121) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1270.966348] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Destroying the VM {{(pid=65121) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1270.966683] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dba2cb76-9dd8-462c-8349-ed9c8cdbbe95 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.975415] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1270.975415] env[65121]: value = "task-5107610" [ 1270.975415] env[65121]: _type = "Task" [ 1270.975415] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.986935] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107610, 'name': Destroy_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.110652] env[65121]: DEBUG nova.scheduler.client.report [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1271.489969] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107610, 'name': Destroy_Task, 'duration_secs': 0.472333} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.489969] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Destroyed the VM [ 1271.489969] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Deleting Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1271.490263] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1dacdb10-3448-4ce2-8f22-8813d0e22839 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.499656] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1271.499656] env[65121]: value = "task-5107612" [ 1271.499656] env[65121]: _type = "Task" [ 1271.499656] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.510612] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107612, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.616199] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.783s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1271.639966] env[65121]: INFO nova.scheduler.client.report [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Deleted allocations for instance b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2 [ 1272.010216] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107612, 'name': RemoveSnapshot_Task, 'duration_secs': 0.375553} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.010651] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Deleted Snapshot of the VM instance {{(pid=65121) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1272.010759] env[65121]: DEBUG nova.compute.manager [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1272.011574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01e2a91-1416-4012-94cb-feb8c431ae1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.148587] env[65121]: DEBUG oslo_concurrency.lockutils [None req-b9cf732e-d66e-448d-9feb-92b93403a10e tempest-AttachVolumeTestJSON-1850153687 tempest-AttachVolumeTestJSON-1850153687-project-member] Lock "b63e4aa2-f8f7-4180-bd2e-d9bfb8d43fb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.504s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1272.527030] env[65121]: INFO nova.compute.manager [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Shelve offloading [ 1273.035756] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1273.036115] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5301062e-ff82-4d6b-ae24-142d9cb88489 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.044996] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1273.044996] env[65121]: value = "task-5107614" [ 1273.044996] env[65121]: _type = "Task" [ 1273.044996] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.056460] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1273.056460] env[65121]: DEBUG nova.compute.manager [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1273.056460] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a574317-5922-4bbe-9185-0577545f30a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.064723] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.064886] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1273.065987] env[65121]: DEBUG nova.network.neutron [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1273.460681] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1273.460864] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993606', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'name': 'volume-50802f92-48fc-45a4-8fdb-0811fade10a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd34b8457-adc1-4d6b-befd-f6f1aff6d1fb', 'attached_at': '', 'detached_at': '', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'serial': '50802f92-48fc-45a4-8fdb-0811fade10a3'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1273.461833] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ec3dfc-0fae-4fb9-87cb-b46389b23236 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.483256] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bb56c9-4845-4467-acdc-f40cae01bd16 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.512507] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] volume-50802f92-48fc-45a4-8fdb-0811fade10a3/volume-50802f92-48fc-45a4-8fdb-0811fade10a3.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1273.512682] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31fb57e1-63db-481b-aa1d-5d1ee7cf501c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.532755] env[65121]: DEBUG oslo_vmware.api [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1273.532755] env[65121]: value = "task-5107615" [ 1273.532755] env[65121]: _type = "Task" [ 1273.532755] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.542103] env[65121]: DEBUG oslo_vmware.api [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107615, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.568415] env[65121]: WARNING neutronclient.v2_0.client [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1273.569872] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1273.570667] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1273.715891] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1273.717269] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1273.787911] env[65121]: WARNING neutronclient.v2_0.client [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1273.788789] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1273.789358] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1273.878420] env[65121]: DEBUG nova.network.neutron [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": "3cbde80d-7df5-4785-821e-b45797716ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1273.996202] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1273.996427] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1274.044150] env[65121]: DEBUG oslo_vmware.api [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107615, 'name': ReconfigVM_Task, 'duration_secs': 0.409044} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.044434] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfigured VM instance instance-0000007a to attach disk [datastore1] volume-50802f92-48fc-45a4-8fdb-0811fade10a3/volume-50802f92-48fc-45a4-8fdb-0811fade10a3.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1274.049666] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb094bd9-5acf-4c65-a2ab-2f981ea41287 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.066520] env[65121]: DEBUG oslo_vmware.api [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1274.066520] env[65121]: value = "task-5107616" [ 1274.066520] env[65121]: _type = "Task" [ 1274.066520] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.074817] env[65121]: DEBUG oslo_vmware.api [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107616, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.381732] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1274.382564] env[65121]: WARNING neutronclient.v2_0.client [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.382954] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.383394] env[65121]: WARNING openstack [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.388485] env[65121]: WARNING neutronclient.v2_0.client [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.498941] env[65121]: DEBUG nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1274.577740] env[65121]: DEBUG oslo_vmware.api [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107616, 'name': ReconfigVM_Task, 'duration_secs': 0.139724} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.578060] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993606', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'name': 'volume-50802f92-48fc-45a4-8fdb-0811fade10a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd34b8457-adc1-4d6b-befd-f6f1aff6d1fb', 'attached_at': '', 'detached_at': '', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'serial': '50802f92-48fc-45a4-8fdb-0811fade10a3'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1274.668961] env[65121]: DEBUG nova.compute.manager [req-519c3c0d-0fc1-4f62-b722-9800effcf162 req-e276443d-38ad-400c-afe1-6e770537e663 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-vif-unplugged-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1274.669145] env[65121]: DEBUG oslo_concurrency.lockutils [req-519c3c0d-0fc1-4f62-b722-9800effcf162 req-e276443d-38ad-400c-afe1-6e770537e663 service nova] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1274.671694] env[65121]: DEBUG oslo_concurrency.lockutils [req-519c3c0d-0fc1-4f62-b722-9800effcf162 req-e276443d-38ad-400c-afe1-6e770537e663 service nova] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1274.671868] env[65121]: DEBUG oslo_concurrency.lockutils [req-519c3c0d-0fc1-4f62-b722-9800effcf162 req-e276443d-38ad-400c-afe1-6e770537e663 service nova] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1274.672051] env[65121]: DEBUG nova.compute.manager [req-519c3c0d-0fc1-4f62-b722-9800effcf162 req-e276443d-38ad-400c-afe1-6e770537e663 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] No waiting events found dispatching network-vif-unplugged-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1274.672242] env[65121]: WARNING nova.compute.manager [req-519c3c0d-0fc1-4f62-b722-9800effcf162 req-e276443d-38ad-400c-afe1-6e770537e663 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received unexpected event network-vif-unplugged-3cbde80d-7df5-4785-821e-b45797716ab3 for instance with vm_state shelved and task_state shelving_offloading. [ 1274.751822] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1274.753011] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751d8218-49b8-4944-8df6-8bfd15211b2f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.761706] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1274.762179] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9cf70cd9-248b-47bb-b2ce-71ae430de235 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.836861] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1274.837236] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1274.837236] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleting the datastore file [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1274.837468] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a17de5fd-316f-47f7-aa59-3f4eaf49e392 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.847024] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1274.847024] env[65121]: value = "task-5107618" [ 1274.847024] env[65121]: _type = "Task" [ 1274.847024] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.857366] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.024439] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1275.024721] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1275.026432] env[65121]: INFO nova.compute.claims [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1275.357031] env[65121]: DEBUG oslo_vmware.api [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187537} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.357398] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1275.357511] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1275.357677] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1275.378078] env[65121]: INFO nova.scheduler.client.report [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted allocations for instance db417aa7-d170-45bc-b2c4-7a97400de1d8 [ 1275.617716] env[65121]: DEBUG nova.objects.instance [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lazy-loading 'flavor' on Instance uuid d34b8457-adc1-4d6b-befd-f6f1aff6d1fb {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1275.882979] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1276.122779] env[65121]: DEBUG oslo_concurrency.lockutils [None req-35b4a8ba-f575-4e18-9095-9082751eb9a5 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.287s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1276.177151] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5e1602-f03f-4576-a78f-d8e1a4603864 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.186153] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4897b4b7-d4f0-4260-9799-6934bfad85ea {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.220076] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d971f7c7-44a3-48b8-aef3-fc5d0fc8ed03 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.228492] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b789e4-c00a-46ba-a9a4-5b55e8dc15ca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.243620] env[65121]: DEBUG nova.compute.provider_tree [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1276.323062] env[65121]: INFO nova.compute.manager [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Rescuing [ 1276.323355] env[65121]: DEBUG oslo_concurrency.lockutils [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1276.323841] env[65121]: DEBUG oslo_concurrency.lockutils [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1276.323841] env[65121]: DEBUG nova.network.neutron [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1276.701417] env[65121]: DEBUG nova.compute.manager [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1276.701764] env[65121]: DEBUG nova.compute.manager [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing instance network info cache due to event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1276.701908] env[65121]: DEBUG oslo_concurrency.lockutils [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Acquiring lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1276.701908] env[65121]: DEBUG oslo_concurrency.lockutils [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Acquired lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1276.702243] env[65121]: DEBUG nova.network.neutron [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1276.747108] env[65121]: DEBUG nova.scheduler.client.report [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1276.826543] env[65121]: WARNING neutronclient.v2_0.client [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1276.827459] env[65121]: WARNING openstack [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1276.827678] env[65121]: WARNING openstack [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1276.969166] env[65121]: WARNING openstack [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1276.969690] env[65121]: WARNING openstack [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.040030] env[65121]: WARNING neutronclient.v2_0.client [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1277.040522] env[65121]: WARNING openstack [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.041083] env[65121]: WARNING openstack [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.127264] env[65121]: DEBUG nova.network.neutron [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1277.205752] env[65121]: WARNING neutronclient.v2_0.client [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1277.206617] env[65121]: WARNING openstack [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.207104] env[65121]: WARNING openstack [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.253245] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1277.254391] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.372s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1277.254615] env[65121]: DEBUG nova.objects.instance [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'resources' on Instance uuid db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.320784] env[65121]: WARNING openstack [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.321252] env[65121]: WARNING openstack [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.364549] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1277.384391] env[65121]: WARNING neutronclient.v2_0.client [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1277.386171] env[65121]: WARNING openstack [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.386812] env[65121]: WARNING openstack [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.473205] env[65121]: DEBUG nova.network.neutron [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updated VIF entry in instance network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1277.473597] env[65121]: DEBUG nova.network.neutron [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1277.631168] env[65121]: DEBUG oslo_concurrency.lockutils [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1277.757062] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "03ce904f-93c2-4635-8c3a-5a54def8be14" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1277.757465] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "03ce904f-93c2-4635-8c3a-5a54def8be14" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1277.759017] env[65121]: DEBUG nova.objects.instance [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'numa_topology' on Instance uuid db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.976172] env[65121]: DEBUG oslo_concurrency.lockutils [req-cb6efc94-96ad-4bce-9192-dc3e8f443d0b req-7a6d0491-d7c9-4dea-8080-92e29b71df17 service nova] Releasing lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1278.261543] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "03ce904f-93c2-4635-8c3a-5a54def8be14" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1278.262139] env[65121]: DEBUG nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1278.265299] env[65121]: DEBUG nova.objects.base [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1278.403912] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a323bd8-1e22-49c8-b07c-2762aed23391 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.412929] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b2b7fc-9185-48f2-9b09-6f0184ebc8a3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.443426] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a359299c-5ffd-4443-b0f3-e66aa48027b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.451814] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ebf1d0-b984-49bc-ab53-8e6935409d6b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.468714] env[65121]: DEBUG nova.compute.provider_tree [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1278.767196] env[65121]: DEBUG nova.compute.utils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1278.768631] env[65121]: DEBUG nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1278.768828] env[65121]: DEBUG nova.network.neutron [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1278.769156] env[65121]: WARNING neutronclient.v2_0.client [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.769460] env[65121]: WARNING neutronclient.v2_0.client [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1278.770100] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1278.770454] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1278.821031] env[65121]: DEBUG nova.policy [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '623fad4bd707488286ffb879396bb7a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cefbef2ba4f0405780211546ebde80fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1278.972553] env[65121]: DEBUG nova.scheduler.client.report [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1279.103271] env[65121]: DEBUG nova.network.neutron [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Successfully created port: a0da2a0e-4940-444a-9fa2-239d9ba589e7 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1279.166719] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1279.167055] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b83e76bf-b23d-4b82-bd23-f9b9f3f6d6d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.175788] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1279.175788] env[65121]: value = "task-5107620" [ 1279.175788] env[65121]: _type = "Task" [ 1279.175788] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.184921] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.278431] env[65121]: DEBUG nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1279.478167] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.224s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1279.686686] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107620, 'name': PowerOffVM_Task, 'duration_secs': 0.187399} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.686951] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1279.687759] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb9c389-4716-400b-a789-c59b9a976b8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.709749] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a625004-9a7d-4da1-8d07-aea1c49e650e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.747431] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1279.747761] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-370705ea-ce45-4787-8656-2cc3e7fdbc7e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.756749] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1279.756749] env[65121]: value = "task-5107622" [ 1279.756749] env[65121]: _type = "Task" [ 1279.756749] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.767867] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] VM already powered off {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1279.768230] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1279.768368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1279.768510] env[65121]: DEBUG oslo_concurrency.lockutils [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1279.768680] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1279.768928] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1536a3b6-d492-4688-9137-8954f9ab772d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.777995] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1279.778170] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1279.779220] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5b0eae6-e145-48d8-a53f-f3bccc2c883e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.789237] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1279.789237] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]5208b5d6-c6df-fd0f-052e-fd93715ed9c0" [ 1279.789237] env[65121]: _type = "Task" [ 1279.789237] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.797486] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5208b5d6-c6df-fd0f-052e-fd93715ed9c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.986751] env[65121]: DEBUG oslo_concurrency.lockutils [None req-7e3dfe1f-300f-4d45-857f-165c9557ce62 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.061s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1279.987507] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.623s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1279.987676] env[65121]: INFO nova.compute.manager [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Unshelving [ 1280.288338] env[65121]: DEBUG nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1280.300215] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]5208b5d6-c6df-fd0f-052e-fd93715ed9c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010182} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.301039] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf690c61-bfcd-4bd9-b2d7-645e181f4a8a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.308045] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1280.308045] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52ffdaf7-021e-3788-e13e-92e796801b43" [ 1280.308045] env[65121]: _type = "Task" [ 1280.308045] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.313077] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1280.313301] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1280.313452] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1280.313621] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1280.313756] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1280.313890] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1280.314100] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1280.314252] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1280.314409] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1280.314556] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1280.314715] env[65121]: DEBUG nova.virt.hardware [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1280.315455] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9298af2-09a7-4935-8ddd-0a01d1e98aa9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.324961] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52ffdaf7-021e-3788-e13e-92e796801b43, 'name': SearchDatastore_Task, 'duration_secs': 0.010105} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.325261] env[65121]: DEBUG oslo_concurrency.lockutils [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1280.325504] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. {{(pid=65121) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1280.326690] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8de6244-e162-4b98-b5bb-0542c3c5bef6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.330488] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e11aca14-af43-4bdc-8b0c-0327f948ac06 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.344104] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1280.344104] env[65121]: value = "task-5107623" [ 1280.344104] env[65121]: _type = "Task" [ 1280.344104] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.352528] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.555847] env[65121]: DEBUG nova.compute.manager [req-0c08a3a1-0b98-4e06-aeed-a3b65e0f53d1 req-ec0e2a6d-d3f5-441b-be8c-fdb4843c0021 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Received event network-vif-plugged-a0da2a0e-4940-444a-9fa2-239d9ba589e7 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1280.556537] env[65121]: DEBUG oslo_concurrency.lockutils [req-0c08a3a1-0b98-4e06-aeed-a3b65e0f53d1 req-ec0e2a6d-d3f5-441b-be8c-fdb4843c0021 service nova] Acquiring lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1280.556767] env[65121]: DEBUG oslo_concurrency.lockutils [req-0c08a3a1-0b98-4e06-aeed-a3b65e0f53d1 req-ec0e2a6d-d3f5-441b-be8c-fdb4843c0021 service nova] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1280.556938] env[65121]: DEBUG oslo_concurrency.lockutils [req-0c08a3a1-0b98-4e06-aeed-a3b65e0f53d1 req-ec0e2a6d-d3f5-441b-be8c-fdb4843c0021 service nova] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1280.557301] env[65121]: DEBUG nova.compute.manager [req-0c08a3a1-0b98-4e06-aeed-a3b65e0f53d1 req-ec0e2a6d-d3f5-441b-be8c-fdb4843c0021 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] No waiting events found dispatching network-vif-plugged-a0da2a0e-4940-444a-9fa2-239d9ba589e7 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1280.557490] env[65121]: WARNING nova.compute.manager [req-0c08a3a1-0b98-4e06-aeed-a3b65e0f53d1 req-ec0e2a6d-d3f5-441b-be8c-fdb4843c0021 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Received unexpected event network-vif-plugged-a0da2a0e-4940-444a-9fa2-239d9ba589e7 for instance with vm_state building and task_state spawning. [ 1280.662832] env[65121]: DEBUG nova.network.neutron [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Successfully updated port: a0da2a0e-4940-444a-9fa2-239d9ba589e7 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1280.853854] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107623, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487429} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.854300] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk. [ 1280.855031] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976d79ad-4eaf-4529-b91b-8c30fd0a866a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.892561] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1280.892910] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78a5b4ae-eaa0-45f4-9edf-524de457a8a4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.912329] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1280.912329] env[65121]: value = "task-5107624" [ 1280.912329] env[65121]: _type = "Task" [ 1280.912329] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.920974] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107624, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.020505] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1281.020836] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1281.021088] env[65121]: DEBUG nova.objects.instance [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'pci_requests' on Instance uuid db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1281.168695] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "refresh_cache-e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.168921] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquired lock "refresh_cache-e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1281.169186] env[65121]: DEBUG nova.network.neutron [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1281.423396] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107624, 'name': ReconfigVM_Task, 'duration_secs': 0.342197} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.423685] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfigured VM instance instance-0000007a to attach disk [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb/3ed5d656-a616-4609-8bb7-b02beb3ac3df-rescue.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1281.424549] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea882860-10a1-4527-b499-8bd8f0ee3e15 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.452931] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31cfd2d9-eb32-4e1a-89b8-3217864310a9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.469284] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1281.469284] env[65121]: value = "task-5107625" [ 1281.469284] env[65121]: _type = "Task" [ 1281.469284] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.477961] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107625, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.525252] env[65121]: DEBUG nova.objects.instance [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'numa_topology' on Instance uuid db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1281.673044] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1281.673398] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1281.711108] env[65121]: DEBUG nova.network.neutron [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1281.731769] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1281.732169] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1281.806538] env[65121]: WARNING neutronclient.v2_0.client [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1281.807225] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1281.807559] env[65121]: WARNING openstack [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1281.954411] env[65121]: DEBUG nova.network.neutron [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Updating instance_info_cache with network_info: [{"id": "a0da2a0e-4940-444a-9fa2-239d9ba589e7", "address": "fa:16:3e:4a:50:4c", "network": {"id": "8d28d91a-6eda-41f4-9692-2d6f7050d087", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1977236942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cefbef2ba4f0405780211546ebde80fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0da2a0e-49", "ovs_interfaceid": "a0da2a0e-4940-444a-9fa2-239d9ba589e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1281.980413] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107625, 'name': ReconfigVM_Task, 'duration_secs': 0.178087} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.980682] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1281.980938] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbc7d473-a660-4e67-a1de-22f8c0a145b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.987978] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1281.987978] env[65121]: value = "task-5107626" [ 1281.987978] env[65121]: _type = "Task" [ 1281.987978] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.996804] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107626, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.029724] env[65121]: INFO nova.compute.claims [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1282.457714] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Releasing lock "refresh_cache-e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1282.458167] env[65121]: DEBUG nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Instance network_info: |[{"id": "a0da2a0e-4940-444a-9fa2-239d9ba589e7", "address": "fa:16:3e:4a:50:4c", "network": {"id": "8d28d91a-6eda-41f4-9692-2d6f7050d087", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1977236942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cefbef2ba4f0405780211546ebde80fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0da2a0e-49", "ovs_interfaceid": "a0da2a0e-4940-444a-9fa2-239d9ba589e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1282.458665] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:50:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae18b41f-e73c-44f1-83dd-467c080944f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0da2a0e-4940-444a-9fa2-239d9ba589e7', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1282.467736] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Creating folder: Project (cefbef2ba4f0405780211546ebde80fb). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1282.468141] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6302c921-fc27-4d80-94d8-e606a55c3cd8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.481161] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Created folder: Project (cefbef2ba4f0405780211546ebde80fb) in parent group-v993268. [ 1282.481365] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Creating folder: Instances. Parent ref: group-v993607. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1282.481608] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8414fac3-e78e-45c0-98d0-cde0607ca66c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.493330] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Created folder: Instances in parent group-v993607. [ 1282.493562] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1282.494151] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1282.494390] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce9e54ee-4291-4f7e-9b33-1d245dc865d6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.514737] env[65121]: DEBUG oslo_vmware.api [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107626, 'name': PowerOnVM_Task, 'duration_secs': 0.424941} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.515488] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1282.518414] env[65121]: DEBUG nova.compute.manager [None req-46d2d56d-6499-40f4-a37d-23150e7a92f9 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1282.519254] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ce4934-dc9a-4c2f-b062-578f37543b61 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.522784] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1282.522784] env[65121]: value = "task-5107629" [ 1282.522784] env[65121]: _type = "Task" [ 1282.522784] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.536893] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107629, 'name': CreateVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.590773] env[65121]: DEBUG nova.compute.manager [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Received event network-changed-a0da2a0e-4940-444a-9fa2-239d9ba589e7 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1282.591037] env[65121]: DEBUG nova.compute.manager [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Refreshing instance network info cache due to event network-changed-a0da2a0e-4940-444a-9fa2-239d9ba589e7. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1282.591486] env[65121]: DEBUG oslo_concurrency.lockutils [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Acquiring lock "refresh_cache-e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.591658] env[65121]: DEBUG oslo_concurrency.lockutils [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Acquired lock "refresh_cache-e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1282.591840] env[65121]: DEBUG nova.network.neutron [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Refreshing network info cache for port a0da2a0e-4940-444a-9fa2-239d9ba589e7 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1283.036101] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107629, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.095861] env[65121]: WARNING neutronclient.v2_0.client [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.096647] env[65121]: WARNING openstack [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.097049] env[65121]: WARNING openstack [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.212472] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f79198c-c58d-4c56-b026-8d2fa5624119 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.216788] env[65121]: WARNING openstack [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.217415] env[65121]: WARNING openstack [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.229517] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cb0af3-4c14-4b3d-b388-26e5c53ad2c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.262124] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a541ce4c-43b7-4097-81b1-e41c2f9ac0ce {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.275160] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1453ab-8a01-4ee3-bf88-f28de608fc52 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.291057] env[65121]: DEBUG nova.compute.provider_tree [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1283.322970] env[65121]: WARNING neutronclient.v2_0.client [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.323524] env[65121]: WARNING openstack [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.323758] env[65121]: WARNING openstack [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.409101] env[65121]: INFO nova.compute.manager [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Unrescuing [ 1283.409491] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.409648] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquired lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1283.409812] env[65121]: DEBUG nova.network.neutron [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1283.498194] env[65121]: DEBUG nova.network.neutron [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Updated VIF entry in instance network info cache for port a0da2a0e-4940-444a-9fa2-239d9ba589e7. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1283.498194] env[65121]: DEBUG nova.network.neutron [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Updating instance_info_cache with network_info: [{"id": "a0da2a0e-4940-444a-9fa2-239d9ba589e7", "address": "fa:16:3e:4a:50:4c", "network": {"id": "8d28d91a-6eda-41f4-9692-2d6f7050d087", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1977236942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cefbef2ba4f0405780211546ebde80fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae18b41f-e73c-44f1-83dd-467c080944f4", "external-id": "nsx-vlan-transportzone-653", "segmentation_id": 653, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0da2a0e-49", "ovs_interfaceid": "a0da2a0e-4940-444a-9fa2-239d9ba589e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1283.538031] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107629, 'name': CreateVM_Task, 'duration_secs': 0.535445} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.538031] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1283.538031] env[65121]: WARNING neutronclient.v2_0.client [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.538031] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.538031] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1283.538031] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1283.538031] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea19a9db-1498-456e-af9e-16a5f4cb3487 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.542126] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1283.542126] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52c8e839-0b32-4ad5-1447-411352b48bc3" [ 1283.542126] env[65121]: _type = "Task" [ 1283.542126] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.551732] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c8e839-0b32-4ad5-1447-411352b48bc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.795235] env[65121]: DEBUG nova.scheduler.client.report [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1283.912705] env[65121]: WARNING neutronclient.v2_0.client [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.913427] env[65121]: WARNING openstack [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.913794] env[65121]: WARNING openstack [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1284.001105] env[65121]: DEBUG oslo_concurrency.lockutils [req-c2915672-ec17-4078-a82e-9c1bbb88d558 req-078d91f5-eda0-4116-a32b-83c20e2b2664 service nova] Releasing lock "refresh_cache-e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.039096] env[65121]: WARNING openstack [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1284.039689] env[65121]: WARNING openstack [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1284.057033] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52c8e839-0b32-4ad5-1447-411352b48bc3, 'name': SearchDatastore_Task, 'duration_secs': 0.010423} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.057033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.057033] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1284.057033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.057033] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1284.057548] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1284.057548] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e3bdbcc-5e56-4140-8ef3-9a479fa2760e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.067424] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1284.067748] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1284.068471] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a60f3f5-a163-4beb-9f67-9a5edfce2988 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.078890] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1284.078890] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524e1fdf-9b40-c627-1f20-907929833e59" [ 1284.078890] env[65121]: _type = "Task" [ 1284.078890] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.087536] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524e1fdf-9b40-c627-1f20-907929833e59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.120499] env[65121]: WARNING neutronclient.v2_0.client [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1284.121643] env[65121]: WARNING openstack [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1284.121643] env[65121]: WARNING openstack [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1284.219162] env[65121]: DEBUG nova.network.neutron [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1284.301603] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.281s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1284.301956] env[65121]: WARNING neutronclient.v2_0.client [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1284.344444] env[65121]: INFO nova.network.neutron [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating port 3cbde80d-7df5-4785-821e-b45797716ab3 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1284.591325] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524e1fdf-9b40-c627-1f20-907929833e59, 'name': SearchDatastore_Task, 'duration_secs': 0.010404} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.592348] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c1ff154-3c2b-4aa9-ad24-8ea79c91ff33 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.598945] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1284.598945] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52bc2831-02e7-93a8-a9c9-120e48a7cd45" [ 1284.598945] env[65121]: _type = "Task" [ 1284.598945] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.609063] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bc2831-02e7-93a8-a9c9-120e48a7cd45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.722493] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Releasing lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.723370] env[65121]: DEBUG nova.objects.instance [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lazy-loading 'flavor' on Instance uuid d34b8457-adc1-4d6b-befd-f6f1aff6d1fb {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1285.109508] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52bc2831-02e7-93a8-a9c9-120e48a7cd45, 'name': SearchDatastore_Task, 'duration_secs': 0.010513} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.109892] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1285.110089] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e/e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1285.110335] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83daf67f-3c1b-4b55-a72d-407706207683 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.117798] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1285.117798] env[65121]: value = "task-5107630" [ 1285.117798] env[65121]: _type = "Task" [ 1285.117798] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.126442] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.230809] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58be17ac-f6ad-406b-adda-b73482cbdb48 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.256487] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1285.256848] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cb03d00-b470-4501-bd45-6de6ca2cc10d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.264848] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1285.264848] env[65121]: value = "task-5107631" [ 1285.264848] env[65121]: _type = "Task" [ 1285.264848] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.275654] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107631, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.628294] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107630, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505122} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.628536] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e/e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1285.628763] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1285.629038] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20fc6ffd-ad9f-43cb-9b67-a6fff253f9a8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.636344] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1285.636344] env[65121]: value = "task-5107632" [ 1285.636344] env[65121]: _type = "Task" [ 1285.636344] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.645171] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107632, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.775493] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107631, 'name': PowerOffVM_Task, 'duration_secs': 0.262137} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.775739] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1285.781184] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfiguring VM instance instance-0000007a to detach disk 2002 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1285.781474] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5f17456-2fbe-4453-9229-fffdc855d0c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.800933] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1285.800933] env[65121]: value = "task-5107633" [ 1285.800933] env[65121]: _type = "Task" [ 1285.800933] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.809713] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107633, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.838394] env[65121]: DEBUG nova.compute.manager [req-4c8d9012-633c-4fe3-b645-bfc300f8ad70 req-74c7c058-1810-4eb8-ab34-f0c405f7567e service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-vif-plugged-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1285.838394] env[65121]: DEBUG oslo_concurrency.lockutils [req-4c8d9012-633c-4fe3-b645-bfc300f8ad70 req-74c7c058-1810-4eb8-ab34-f0c405f7567e service nova] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1285.838696] env[65121]: DEBUG oslo_concurrency.lockutils [req-4c8d9012-633c-4fe3-b645-bfc300f8ad70 req-74c7c058-1810-4eb8-ab34-f0c405f7567e service nova] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1285.838696] env[65121]: DEBUG oslo_concurrency.lockutils [req-4c8d9012-633c-4fe3-b645-bfc300f8ad70 req-74c7c058-1810-4eb8-ab34-f0c405f7567e service nova] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1285.838828] env[65121]: DEBUG nova.compute.manager [req-4c8d9012-633c-4fe3-b645-bfc300f8ad70 req-74c7c058-1810-4eb8-ab34-f0c405f7567e service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] No waiting events found dispatching network-vif-plugged-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1285.838939] env[65121]: WARNING nova.compute.manager [req-4c8d9012-633c-4fe3-b645-bfc300f8ad70 req-74c7c058-1810-4eb8-ab34-f0c405f7567e service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received unexpected event network-vif-plugged-3cbde80d-7df5-4785-821e-b45797716ab3 for instance with vm_state shelved_offloaded and task_state spawning. [ 1285.927978] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.927978] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1285.927978] env[65121]: DEBUG nova.network.neutron [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1286.147998] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107632, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070077} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.148412] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1286.149262] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e67c1c5-f4e8-41f8-8606-3a9561f35bbe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.175621] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Reconfiguring VM instance instance-0000007e to attach disk [datastore1] e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e/e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1286.175950] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e79c47a-5cd3-4627-8223-e5ab416e514d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.196637] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1286.196637] env[65121]: value = "task-5107634" [ 1286.196637] env[65121]: _type = "Task" [ 1286.196637] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.205573] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107634, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.312244] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107633, 'name': ReconfigVM_Task, 'duration_secs': 0.379585} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.312554] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfigured VM instance instance-0000007a to detach disk 2002 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1286.312736] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1286.313028] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b5223a6-3cb2-42b6-93d3-ac0b911543ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.319816] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1286.319816] env[65121]: value = "task-5107635" [ 1286.319816] env[65121]: _type = "Task" [ 1286.319816] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.328406] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.430423] env[65121]: WARNING neutronclient.v2_0.client [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1286.431145] env[65121]: WARNING openstack [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1286.431494] env[65121]: WARNING openstack [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.573161] env[65121]: WARNING openstack [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1286.573595] env[65121]: WARNING openstack [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.652103] env[65121]: WARNING neutronclient.v2_0.client [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1286.653195] env[65121]: WARNING openstack [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1286.653584] env[65121]: WARNING openstack [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.711014] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107634, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.760272] env[65121]: DEBUG nova.network.neutron [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": "3cbde80d-7df5-4785-821e-b45797716ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1286.830241] env[65121]: DEBUG oslo_vmware.api [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107635, 'name': PowerOnVM_Task, 'duration_secs': 0.468869} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.830510] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1286.830742] env[65121]: DEBUG nova.compute.manager [None req-f19ceb8a-92dc-4393-b061-41a3eccab503 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1286.831549] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1842a89a-f7ac-4f3e-8a86-054e2c3adbb5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.942361] env[65121]: DEBUG oslo_concurrency.lockutils [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "b669fb5d-a68f-4390-b903-80879c239b4e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1286.942779] env[65121]: DEBUG oslo_concurrency.lockutils [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1286.943032] env[65121]: DEBUG nova.compute.manager [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1286.944169] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f10487-0432-4447-b6f2-c781d7cc7f90 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.953737] env[65121]: DEBUG nova.compute.manager [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65121) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1286.954486] env[65121]: DEBUG nova.objects.instance [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'flavor' on Instance uuid b669fb5d-a68f-4390-b903-80879c239b4e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1287.210024] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107634, 'name': ReconfigVM_Task, 'duration_secs': 0.532007} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.210434] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Reconfigured VM instance instance-0000007e to attach disk [datastore1] e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e/e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1287.210853] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3a1c9fc-355e-48e8-a1a1-d3ab5ba1e0c7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.218400] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1287.218400] env[65121]: value = "task-5107636" [ 1287.218400] env[65121]: _type = "Task" [ 1287.218400] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.227197] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107636, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.263533] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1287.294542] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6076e395aae700e94e8648fd850b21c9',container_format='bare',created_at=2025-12-12T14:31:24Z,direct_url=,disk_format='vmdk',id=b118bbe9-8659-4100-a942-549907f6e6cf,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-2138370773-shelved',owner='4467da3ed41245ddbc93fc865a8b7bdd',properties=ImageMetaProps,protected=,size=31662592,status='active',tags=,updated_at=2025-12-12T14:31:39Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1287.294800] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1287.295000] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1287.295153] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1287.295284] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1287.295424] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1287.295626] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1287.295773] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1287.295929] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1287.296094] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1287.296334] env[65121]: DEBUG nova.virt.hardware [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1287.297285] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329fed03-0ff5-40fc-af1b-bf05aa3094bc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.306671] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f5bb9d-830b-4fa3-a3bf-267a939bd0d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.322517] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:17:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cbde80d-7df5-4785-821e-b45797716ab3', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1287.330016] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1287.330317] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1287.330546] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a8ad926-39d7-47e5-8a1e-3e7ed10acbb2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.352939] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1287.352939] env[65121]: value = "task-5107637" [ 1287.352939] env[65121]: _type = "Task" [ 1287.352939] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.361742] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107637, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.729575] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107636, 'name': Rename_Task, 'duration_secs': 0.283276} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.729836] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1287.730156] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76045c21-0e17-4c73-8d63-1fa6c857d9f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.740330] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1287.740330] env[65121]: value = "task-5107638" [ 1287.740330] env[65121]: _type = "Task" [ 1287.740330] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.750249] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107638, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.868237] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107637, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.879286] env[65121]: DEBUG nova.compute.manager [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1287.879443] env[65121]: DEBUG nova.compute.manager [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing instance network info cache due to event network-changed-3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1287.879654] env[65121]: DEBUG oslo_concurrency.lockutils [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Acquiring lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.879792] env[65121]: DEBUG oslo_concurrency.lockutils [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Acquired lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1287.879946] env[65121]: DEBUG nova.network.neutron [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Refreshing network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1287.966370] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1287.966558] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca80b70d-b3b3-4277-ae73-7570dca3d1a0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.976526] env[65121]: DEBUG oslo_vmware.api [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1287.976526] env[65121]: value = "task-5107639" [ 1287.976526] env[65121]: _type = "Task" [ 1287.976526] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.987641] env[65121]: DEBUG oslo_vmware.api [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.251762] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107638, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.365450] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107637, 'name': CreateVM_Task, 'duration_secs': 0.524526} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.365655] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1288.366393] env[65121]: WARNING neutronclient.v2_0.client [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1288.366852] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.367069] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1288.367663] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1288.367964] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82bdcb46-099f-4cc7-8d48-43bc7407c0f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.374378] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1288.374378] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b0827-c4e1-786b-150d-0eb9d1443fe9" [ 1288.374378] env[65121]: _type = "Task" [ 1288.374378] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.385578] env[65121]: WARNING neutronclient.v2_0.client [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1288.386246] env[65121]: WARNING openstack [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.386582] env[65121]: WARNING openstack [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.393526] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524b0827-c4e1-786b-150d-0eb9d1443fe9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.418502] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.418889] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.488776] env[65121]: DEBUG oslo_vmware.api [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107639, 'name': PowerOffVM_Task, 'duration_secs': 0.20541} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.489118] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1288.489372] env[65121]: DEBUG nova.compute.manager [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1288.490351] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7017755f-c747-40e6-9c8c-13517809e7cf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.519092] env[65121]: WARNING openstack [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.519235] env[65121]: WARNING openstack [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.587427] env[65121]: WARNING neutronclient.v2_0.client [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1288.588236] env[65121]: WARNING openstack [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1288.588681] env[65121]: WARNING openstack [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1288.673741] env[65121]: DEBUG nova.network.neutron [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updated VIF entry in instance network info cache for port 3cbde80d-7df5-4785-821e-b45797716ab3. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1288.674168] env[65121]: DEBUG nova.network.neutron [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [{"id": "3cbde80d-7df5-4785-821e-b45797716ab3", "address": "fa:16:3e:96:17:82", "network": {"id": "f943d5ed-7eb8-40d8-a7ad-f56f082c562b", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1222794930-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4467da3ed41245ddbc93fc865a8b7bdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cbde80d-7d", "ovs_interfaceid": "3cbde80d-7df5-4785-821e-b45797716ab3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1288.751543] env[65121]: DEBUG oslo_vmware.api [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107638, 'name': PowerOnVM_Task, 'duration_secs': 0.735342} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.751845] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1288.752555] env[65121]: INFO nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Took 8.46 seconds to spawn the instance on the hypervisor. [ 1288.752555] env[65121]: DEBUG nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1288.753212] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58b431d-bf4a-46d1-be6e-90ecc74457d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.888084] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1288.888084] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Processing image b118bbe9-8659-4100-a942-549907f6e6cf {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1288.888391] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf/b118bbe9-8659-4100-a942-549907f6e6cf.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.888513] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf/b118bbe9-8659-4100-a942-549907f6e6cf.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1288.888807] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1288.889229] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd912cfc-2407-4f9d-9732-f928c5262a0f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.903434] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1288.903434] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1288.903434] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dc67b94-3907-4d08-b578-47de1edfeb82 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.910436] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1288.910436] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]529a7229-46bd-ae47-0cdd-f5a97e0c59c4" [ 1288.910436] env[65121]: _type = "Task" [ 1288.910436] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.920639] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]529a7229-46bd-ae47-0cdd-f5a97e0c59c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.926724] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.926971] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.927874] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.927874] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.927874] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.927874] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.927874] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1288.928497] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.004574] env[65121]: DEBUG oslo_concurrency.lockutils [None req-724925e7-1755-4368-a828-307b5aa8f9ac tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.061s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1289.177048] env[65121]: DEBUG oslo_concurrency.lockutils [req-e85dfadb-6ae7-4e72-86f5-09fc60eae4d3 req-55c0e15c-61f4-41c0-8b0f-252f7c9c283b service nova] Releasing lock "refresh_cache-db417aa7-d170-45bc-b2c4-7a97400de1d8" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1289.272172] env[65121]: INFO nova.compute.manager [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Took 14.27 seconds to build instance. [ 1289.404632] env[65121]: DEBUG nova.objects.instance [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'flavor' on Instance uuid b669fb5d-a68f-4390-b903-80879c239b4e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1289.422417] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Preparing fetch location {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1289.422658] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Fetch image to [datastore2] OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14/OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14.vmdk {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1289.422825] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Downloading stream optimized image b118bbe9-8659-4100-a942-549907f6e6cf to [datastore2] OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14/OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14.vmdk on the data store datastore2 as vApp {{(pid=65121) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1289.422984] env[65121]: DEBUG nova.virt.vmwareapi.images [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Downloading image file data b118bbe9-8659-4100-a942-549907f6e6cf to the ESX as VM named 'OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14' {{(pid=65121) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1289.431715] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1289.431946] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1289.432119] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1289.432265] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1289.433274] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c2982a-e84a-4bd6-92ef-b41f1e95b3d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.443060] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b249b8-1401-4953-8ff3-c574c7fc2470 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.494084] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e61793-a462-4181-b8e2-ab3bef8da608 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.505934] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4694047b-ba82-41ab-9f56-d49d6c113a31 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.561476] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178493MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1289.561706] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1289.562034] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1289.586890] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1289.586890] env[65121]: value = "resgroup-9" [ 1289.586890] env[65121]: _type = "ResourcePool" [ 1289.586890] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1289.587161] env[65121]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-817e31ba-685f-4783-8c8a-1a27a0292e99 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.609972] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease: (returnval){ [ 1289.609972] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b95c1-2033-c6de-72ff-50ad33ac6b1c" [ 1289.609972] env[65121]: _type = "HttpNfcLease" [ 1289.609972] env[65121]: } obtained for vApp import into resource pool (val){ [ 1289.609972] env[65121]: value = "resgroup-9" [ 1289.609972] env[65121]: _type = "ResourcePool" [ 1289.609972] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1289.610351] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the lease: (returnval){ [ 1289.610351] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b95c1-2033-c6de-72ff-50ad33ac6b1c" [ 1289.610351] env[65121]: _type = "HttpNfcLease" [ 1289.610351] env[65121]: } to be ready. {{(pid=65121) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1289.619465] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1289.619465] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b95c1-2033-c6de-72ff-50ad33ac6b1c" [ 1289.619465] env[65121]: _type = "HttpNfcLease" [ 1289.619465] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1289.634925] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1289.773975] env[65121]: DEBUG oslo_concurrency.lockutils [None req-03a3eb65-a0e7-4858-bd9a-2edb27ae96db tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.777s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1289.774377] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.140s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1289.774614] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1289.774837] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1289.775021] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1289.777398] env[65121]: INFO nova.compute.manager [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Terminating instance [ 1289.909809] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.909982] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1289.910200] env[65121]: DEBUG nova.network.neutron [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1289.910384] env[65121]: DEBUG nova.objects.instance [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'info_cache' on Instance uuid b669fb5d-a68f-4390-b903-80879c239b4e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1289.921743] env[65121]: DEBUG nova.compute.manager [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Received event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1289.922085] env[65121]: DEBUG nova.compute.manager [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing instance network info cache due to event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1289.922266] env[65121]: DEBUG oslo_concurrency.lockutils [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Acquiring lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.922336] env[65121]: DEBUG oslo_concurrency.lockutils [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Acquired lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1289.922492] env[65121]: DEBUG nova.network.neutron [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1290.121355] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1290.121355] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b95c1-2033-c6de-72ff-50ad33ac6b1c" [ 1290.121355] env[65121]: _type = "HttpNfcLease" [ 1290.121355] env[65121]: } is initializing. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1290.281443] env[65121]: DEBUG nova.compute.manager [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1290.281742] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1290.282659] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd5ab37-d64d-47e6-ac7d-ef6d202d32bb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.291451] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1290.291785] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6c81bac-7c80-43bd-ad1c-9a46c604fa69 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.300291] env[65121]: DEBUG oslo_vmware.api [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1290.300291] env[65121]: value = "task-5107641" [ 1290.300291] env[65121]: _type = "Task" [ 1290.300291] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.309349] env[65121]: DEBUG oslo_vmware.api [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107641, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.413813] env[65121]: DEBUG nova.objects.base [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65121) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1290.427743] env[65121]: WARNING neutronclient.v2_0.client [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1290.428530] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1290.428873] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1290.550513] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1290.550908] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1290.595933] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 4a72fedd-b114-468e-8f34-0caec6ce73fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596102] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 3859dd02-9b52-4c71-a41f-c41f47ac1706 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596226] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance d49a301c-c0d5-476e-9f1f-b8f98cf6adb7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596340] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 992ba328-206a-4d47-a28b-f3ef9b20817f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596451] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance d34b8457-adc1-4d6b-befd-f6f1aff6d1fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596573] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 999ab257-c116-4a73-9b87-b52aeb9a3bb3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596683] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance b669fb5d-a68f-4390-b903-80879c239b4e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596789] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 91a36110-5e97-4553-90c3-3e4213e30c14 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596893] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance 07ff0535-f83c-4a1b-8077-26a8b1fa02b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.596996] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.597128] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Instance db417aa7-d170-45bc-b2c4-7a97400de1d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65121) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.597318] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1290.597463] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=100GB used_disk=10GB total_vcpus=48 used_vcpus=11 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '11', 'num_vm_active': '8', 'num_task_None': '8', 'num_os_type_None': '11', 'num_proj_00383aa3355e438cb703c2b86c7917f2': '3', 'io_workload': '0', 'num_proj_ad1aea30d62c45e193c5a54c429ce7ba': '1', 'num_proj_a23c876483f34b1db8beec9a669f7822': '3', 'num_vm_rescued': '1', 'num_vm_shelved_offloaded': '1', 'num_task_spawning': '1', 'num_proj_4467da3ed41245ddbc93fc865a8b7bdd': '1', 'num_vm_stopped': '1', 'num_task_powering-on': '1', 'num_proj_72237a0d762645588c41231b0a34a796': '1', 'num_proj_f056059180af48bba4587006efb0b211': '1', 'num_task_deleting': '1', 'num_proj_cefbef2ba4f0405780211546ebde80fb': '1'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1290.617107] env[65121]: WARNING neutronclient.v2_0.client [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1290.617766] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1290.618112] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1290.637842] env[65121]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1290.637842] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b95c1-2033-c6de-72ff-50ad33ac6b1c" [ 1290.637842] env[65121]: _type = "HttpNfcLease" [ 1290.637842] env[65121]: } is ready. {{(pid=65121) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1290.638231] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1290.638231] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524b95c1-2033-c6de-72ff-50ad33ac6b1c" [ 1290.638231] env[65121]: _type = "HttpNfcLease" [ 1290.638231] env[65121]: }. {{(pid=65121) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1290.638979] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb429455-72f4-4311-873d-bf6c0d90221a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.650470] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222967c-17fc-b897-8c76-1b2b0c8302a3/disk-0.vmdk from lease info. {{(pid=65121) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1290.650774] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating HTTP connection to write to file with size = 31662592 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222967c-17fc-b897-8c76-1b2b0c8302a3/disk-0.vmdk. {{(pid=65121) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1290.718679] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5d7543ba-d7ae-434b-9a9f-4f43095657d5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.758733] env[65121]: DEBUG nova.network.neutron [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updated VIF entry in instance network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1290.759181] env[65121]: DEBUG nova.network.neutron [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1290.813650] env[65121]: DEBUG oslo_vmware.api [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107641, 'name': PowerOffVM_Task, 'duration_secs': 0.391587} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.813921] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1290.814094] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1290.814361] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4c3693b6-f734-45c8-8c92-ff4114dde60e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.819279] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77451ec3-4189-4791-9957-e77542820225 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.828228] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5207d693-360c-4f50-898e-0ede11063dfe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.862261] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8eb708-b8fe-41e6-b82c-e785a54ff142 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.872341] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab48d17-76ed-41b5-9a88-6eae586b7426 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.888608] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1290.891481] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1290.891685] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1290.891857] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Deleting the datastore file [datastore1] e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1290.892433] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82d7d9d9-f573-4991-b66a-266f5c68d24a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.900290] env[65121]: DEBUG oslo_vmware.api [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for the task: (returnval){ [ 1290.900290] env[65121]: value = "task-5107643" [ 1290.900290] env[65121]: _type = "Task" [ 1290.900290] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.910679] env[65121]: DEBUG oslo_vmware.api [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107643, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.916921] env[65121]: WARNING neutronclient.v2_0.client [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1290.917549] env[65121]: WARNING openstack [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1290.918119] env[65121]: WARNING openstack [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.039021] env[65121]: WARNING openstack [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.039681] env[65121]: WARNING openstack [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.107338] env[65121]: WARNING neutronclient.v2_0.client [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.108157] env[65121]: WARNING openstack [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.108579] env[65121]: WARNING openstack [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.211865] env[65121]: DEBUG nova.network.neutron [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updating instance_info_cache with network_info: [{"id": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "address": "fa:16:3e:2b:98:34", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d86d6-62", "ovs_interfaceid": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1291.263121] env[65121]: DEBUG oslo_concurrency.lockutils [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Releasing lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1291.263121] env[65121]: DEBUG nova.compute.manager [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Received event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1291.263121] env[65121]: DEBUG nova.compute.manager [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing instance network info cache due to event network-changed-1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1291.263121] env[65121]: DEBUG oslo_concurrency.lockutils [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Acquiring lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1291.263121] env[65121]: DEBUG oslo_concurrency.lockutils [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Acquired lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1291.263121] env[65121]: DEBUG nova.network.neutron [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Refreshing network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1291.394037] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1291.415211] env[65121]: DEBUG oslo_vmware.api [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Task: {'id': task-5107643, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134811} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.418444] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1291.418749] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1291.418883] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1291.419471] env[65121]: INFO nova.compute.manager [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1291.419471] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1291.419923] env[65121]: DEBUG nova.compute.manager [-] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1291.420076] env[65121]: DEBUG nova.network.neutron [-] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1291.420360] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.420933] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.421379] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.464445] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.717614] env[65121]: DEBUG oslo_concurrency.lockutils [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1291.731376] env[65121]: DEBUG nova.compute.manager [req-5c74f48c-583d-4de1-8da7-3400c367822f req-5f588500-e833-4ef7-bcbe-52584e1ed883 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Received event network-vif-deleted-a0da2a0e-4940-444a-9fa2-239d9ba589e7 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1291.731609] env[65121]: INFO nova.compute.manager [req-5c74f48c-583d-4de1-8da7-3400c367822f req-5f588500-e833-4ef7-bcbe-52584e1ed883 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Neutron deleted interface a0da2a0e-4940-444a-9fa2-239d9ba589e7; detaching it from the instance and deleting it from the info cache [ 1291.731796] env[65121]: DEBUG nova.network.neutron [req-5c74f48c-583d-4de1-8da7-3400c367822f req-5f588500-e833-4ef7-bcbe-52584e1ed883 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1291.767831] env[65121]: WARNING neutronclient.v2_0.client [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.768806] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.769295] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.902762] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.903162] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.909927] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1291.910108] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.348s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1291.969038] env[65121]: WARNING neutronclient.v2_0.client [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.970578] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.971097] env[65121]: WARNING openstack [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1292.077221] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Completed reading data from the image iterator. {{(pid=65121) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1292.077679] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222967c-17fc-b897-8c76-1b2b0c8302a3/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1292.078880] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf6fd39-6ca9-48d8-ba13-54f57078ec71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.084010] env[65121]: DEBUG nova.network.neutron [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updated VIF entry in instance network info cache for port 1c9a1903-516b-4056-ba71-cf778ceebf76. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1292.084454] env[65121]: DEBUG nova.network.neutron [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [{"id": "1c9a1903-516b-4056-ba71-cf778ceebf76", "address": "fa:16:3e:4f:99:80", "network": {"id": "4004c564-5803-461d-8d8b-018c3fad1f96", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-484766091-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a23c876483f34b1db8beec9a669f7822", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "12d8eedb-97cb-4d3b-b364-42d7fd8b3c85", "external-id": "nsx-vlan-transportzone-870", "segmentation_id": 870, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c9a1903-51", "ovs_interfaceid": "1c9a1903-516b-4056-ba71-cf778ceebf76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1292.089871] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222967c-17fc-b897-8c76-1b2b0c8302a3/disk-0.vmdk is in state: ready. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1292.090088] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222967c-17fc-b897-8c76-1b2b0c8302a3/disk-0.vmdk. {{(pid=65121) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1292.090321] env[65121]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-65d1eeb0-129d-410c-8172-8767a73836de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.200435] env[65121]: DEBUG nova.network.neutron [-] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1292.234691] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8128cd7e-2ab0-4e00-8377-959feaf04842 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.249564] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9a2381-55fa-4cc3-9820-e0dee318bf08 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.299209] env[65121]: DEBUG nova.compute.manager [req-5c74f48c-583d-4de1-8da7-3400c367822f req-5f588500-e833-4ef7-bcbe-52584e1ed883 service nova] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Detach interface failed, port_id=a0da2a0e-4940-444a-9fa2-239d9ba589e7, reason: Instance e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1292.548080] env[65121]: DEBUG oslo_vmware.rw_handles [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5222967c-17fc-b897-8c76-1b2b0c8302a3/disk-0.vmdk. {{(pid=65121) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1292.548080] env[65121]: INFO nova.virt.vmwareapi.images [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Downloaded image file data b118bbe9-8659-4100-a942-549907f6e6cf [ 1292.548875] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef513191-cec9-4750-82bd-b9e9cc71093a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.566171] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8217e44b-404c-4c6d-9683-bd0e0043b049 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.587157] env[65121]: DEBUG oslo_concurrency.lockutils [req-0f416ec4-c5d8-4105-b609-10ead9656251 req-e0b58148-90e4-4912-a889-e522df2010f4 service nova] Releasing lock "refresh_cache-d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1292.703306] env[65121]: INFO nova.compute.manager [-] [instance: e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e] Took 1.28 seconds to deallocate network for instance. [ 1292.723799] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1292.724133] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d728127-96be-413a-93ed-7b2db1f64329 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.732429] env[65121]: DEBUG oslo_vmware.api [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1292.732429] env[65121]: value = "task-5107645" [ 1292.732429] env[65121]: _type = "Task" [ 1292.732429] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.741896] env[65121]: DEBUG oslo_vmware.api [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107645, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.743481] env[65121]: INFO nova.virt.vmwareapi.images [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] The imported VM was unregistered [ 1292.746691] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Caching image {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1292.746962] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Creating directory with path [datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1292.747222] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13c4d5eb-bf50-425c-8833-763ba52a65df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.774095] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Created directory with path [datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1292.774358] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14/OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14.vmdk to [datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf/b118bbe9-8659-4100-a942-549907f6e6cf.vmdk. {{(pid=65121) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1292.774655] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-49ea899f-5eee-435e-a22e-ac5e3cbc74fc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.783215] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1292.783215] env[65121]: value = "task-5107646" [ 1292.783215] env[65121]: _type = "Task" [ 1292.783215] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.792186] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107646, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.211742] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1293.212345] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1293.212761] env[65121]: DEBUG nova.objects.instance [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lazy-loading 'resources' on Instance uuid e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1293.247522] env[65121]: DEBUG oslo_vmware.api [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107645, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.298545] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107646, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.746506] env[65121]: DEBUG oslo_vmware.api [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107645, 'name': PowerOnVM_Task, 'duration_secs': 0.536359} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.746950] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1293.746950] env[65121]: DEBUG nova.compute.manager [None req-4135813d-f5ba-4d28-a07d-8f724428b5a2 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1293.747648] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d51f746-edb9-4e37-b387-e231911931ed {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.795951] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107646, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.879778] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ccda0a-13b4-4cc6-938a-8eb3b43d6b0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.892926] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32ee323-6659-4fb9-a16e-e24259fe0c64 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.930196] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134e3c19-8e78-44b3-88fe-c80c034a9c04 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.943063] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9ac456-9494-4cd1-a15b-cedc5ce57887 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.962929] env[65121]: DEBUG nova.compute.provider_tree [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1294.299276] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107646, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.466352] env[65121]: DEBUG nova.scheduler.client.report [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1294.798173] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107646, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.954394] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1294.955170] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1294.955170] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1294.955170] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1294.955170] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1294.957833] env[65121]: INFO nova.compute.manager [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Terminating instance [ 1294.971608] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1295.000470] env[65121]: INFO nova.scheduler.client.report [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Deleted allocations for instance e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e [ 1295.298884] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107646, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.451949] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c83ed91-63bd-48c8-b57f-462bb0e52863 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.458919] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Suspending the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1295.459203] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-b720aa9a-1d61-4619-bfe3-011051af7eae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.463529] env[65121]: DEBUG nova.compute.manager [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1295.463747] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1295.464574] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387825c4-71cc-4760-891f-133c60b0222d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.469174] env[65121]: DEBUG oslo_vmware.api [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1295.469174] env[65121]: value = "task-5107647" [ 1295.469174] env[65121]: _type = "Task" [ 1295.469174] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.475250] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1295.475853] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03ddb25a-1f24-42cc-b422-e2091c8d4571 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.481611] env[65121]: DEBUG oslo_vmware.api [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107647, 'name': SuspendVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.486322] env[65121]: DEBUG oslo_vmware.api [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1295.486322] env[65121]: value = "task-5107648" [ 1295.486322] env[65121]: _type = "Task" [ 1295.486322] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.495134] env[65121]: DEBUG oslo_vmware.api [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107648, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.508790] env[65121]: DEBUG oslo_concurrency.lockutils [None req-43392120-10d4-41ef-9a48-fee9ffc4ece9 tempest-ServerGroupTestJSON-760400203 tempest-ServerGroupTestJSON-760400203-project-member] Lock "e2d8477f-bdf4-4ee6-9dbf-3e54da60be6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.734s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1295.797143] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107646, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.611866} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.797433] env[65121]: INFO nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14/OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14.vmdk to [datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf/b118bbe9-8659-4100-a942-549907f6e6cf.vmdk. [ 1295.797615] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Cleaning up location [datastore2] OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14 {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1295.797771] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_4668a085-82be-4a32-8381-f086110e7d14 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1295.798037] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b04e6e7f-a0ca-46ec-a2e1-e45d232550b3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.806200] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1295.806200] env[65121]: value = "task-5107649" [ 1295.806200] env[65121]: _type = "Task" [ 1295.806200] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.816515] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.980449] env[65121]: DEBUG oslo_vmware.api [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107647, 'name': SuspendVM_Task} progress is 12%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.997161] env[65121]: DEBUG oslo_vmware.api [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107648, 'name': PowerOffVM_Task, 'duration_secs': 0.211139} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.997497] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1295.997671] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1295.997939] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-002e4093-9cad-4e1e-9bc0-984b81376766 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.073903] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1296.074224] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1296.074419] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleting the datastore file [datastore1] 999ab257-c116-4a73-9b87-b52aeb9a3bb3 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1296.074810] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc6f8307-e7e2-436c-82cb-6901b7436160 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.083301] env[65121]: DEBUG oslo_vmware.api [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for the task: (returnval){ [ 1296.083301] env[65121]: value = "task-5107651" [ 1296.083301] env[65121]: _type = "Task" [ 1296.083301] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.094550] env[65121]: DEBUG oslo_vmware.api [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.315877] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045988} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.316165] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1296.316327] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf/b118bbe9-8659-4100-a942-549907f6e6cf.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1296.316566] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf/b118bbe9-8659-4100-a942-549907f6e6cf.vmdk to [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1296.316827] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43abd204-2fbc-4c9b-bace-c88c8b898c71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.324996] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1296.324996] env[65121]: value = "task-5107652" [ 1296.324996] env[65121]: _type = "Task" [ 1296.324996] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.333940] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.482387] env[65121]: DEBUG oslo_vmware.api [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107647, 'name': SuspendVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.597268] env[65121]: DEBUG oslo_vmware.api [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Task: {'id': task-5107651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200452} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.597789] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1296.598150] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1296.598355] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1296.598530] env[65121]: INFO nova.compute.manager [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1296.598936] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1296.599315] env[65121]: DEBUG nova.compute.manager [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1296.599503] env[65121]: DEBUG nova.network.neutron [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1296.599867] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1296.600821] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1296.601210] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1296.678073] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1296.836996] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107652, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.987915] env[65121]: DEBUG oslo_vmware.api [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107647, 'name': SuspendVM_Task, 'duration_secs': 1.094373} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.988277] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Suspended the VM {{(pid=65121) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1296.988462] env[65121]: DEBUG nova.compute.manager [None req-6b85e6eb-157e-49ef-b1fd-9f9ec4c5fcc0 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1296.989316] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda88ee1-4529-424b-b32f-2ff87a2395d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.205226] env[65121]: DEBUG nova.compute.manager [req-1982f7fc-3b4f-46b0-b20f-10875ce44b5a req-d5e1b2c8-9d1f-4861-95f2-4cf48803b812 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Received event network-vif-deleted-91969a04-4a8e-4334-829f-3e3ae6828511 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1297.205764] env[65121]: INFO nova.compute.manager [req-1982f7fc-3b4f-46b0-b20f-10875ce44b5a req-d5e1b2c8-9d1f-4861-95f2-4cf48803b812 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Neutron deleted interface 91969a04-4a8e-4334-829f-3e3ae6828511; detaching it from the instance and deleting it from the info cache [ 1297.206190] env[65121]: DEBUG nova.network.neutron [req-1982f7fc-3b4f-46b0-b20f-10875ce44b5a req-d5e1b2c8-9d1f-4861-95f2-4cf48803b812 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1297.344885] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107652, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.654068] env[65121]: DEBUG nova.network.neutron [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1297.713856] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1405296c-05d1-4d16-9eaa-2d7f7ba76938 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.725478] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b5cc7e-f62a-4435-ae61-1ae1af9f7bb1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.764949] env[65121]: DEBUG nova.compute.manager [req-1982f7fc-3b4f-46b0-b20f-10875ce44b5a req-d5e1b2c8-9d1f-4861-95f2-4cf48803b812 service nova] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Detach interface failed, port_id=91969a04-4a8e-4334-829f-3e3ae6828511, reason: Instance 999ab257-c116-4a73-9b87-b52aeb9a3bb3 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1297.838474] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107652, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.892311] env[65121]: INFO nova.compute.manager [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Resuming [ 1297.893096] env[65121]: DEBUG nova.objects.instance [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'flavor' on Instance uuid b669fb5d-a68f-4390-b903-80879c239b4e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1298.156427] env[65121]: INFO nova.compute.manager [-] [instance: 999ab257-c116-4a73-9b87-b52aeb9a3bb3] Took 1.56 seconds to deallocate network for instance. [ 1298.339533] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107652, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.664076] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1298.664356] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1298.664582] env[65121]: DEBUG nova.objects.instance [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lazy-loading 'resources' on Instance uuid 999ab257-c116-4a73-9b87-b52aeb9a3bb3 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1298.840602] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107652, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.476479} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.840981] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b118bbe9-8659-4100-a942-549907f6e6cf/b118bbe9-8659-4100-a942-549907f6e6cf.vmdk to [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1298.841842] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b19142a-6f97-4ed1-a2f1-82421242bd21 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.865240] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1298.865661] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f158d6e-af6d-4969-b44b-0b7eb63a08eb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.888438] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1298.888438] env[65121]: value = "task-5107653" [ 1298.888438] env[65121]: _type = "Task" [ 1298.888438] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.899756] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107653, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.312970] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0790e63a-3eb1-4972-91d7-d618ed4568de {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.321447] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef79acb9-4217-42fd-ac1f-13329011d4d8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.352109] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d37342f-f7c5-48d3-8249-6a7c2c611a84 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.360202] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158ae69a-c87f-4683-995b-47e73e9f9982 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.374355] env[65121]: DEBUG nova.compute.provider_tree [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1299.399976] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107653, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.406972] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.407146] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquired lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1299.407316] env[65121]: DEBUG nova.network.neutron [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1299.877809] env[65121]: DEBUG nova.scheduler.client.report [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1299.901132] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107653, 'name': ReconfigVM_Task, 'duration_secs': 0.790498} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.901449] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Reconfigured VM instance instance-00000078 to attach disk [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8/db417aa7-d170-45bc-b2c4-7a97400de1d8.vmdk or device None with type streamOptimized {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1299.902194] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bbedbd4d-b8c2-4b0d-a4a4-c98532a8bbf9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.908773] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1299.908773] env[65121]: value = "task-5107654" [ 1299.908773] env[65121]: _type = "Task" [ 1299.908773] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.909369] env[65121]: WARNING neutronclient.v2_0.client [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1299.910015] env[65121]: WARNING openstack [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1299.910412] env[65121]: WARNING openstack [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1299.928412] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107654, 'name': Rename_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.027023] env[65121]: WARNING openstack [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1300.027480] env[65121]: WARNING openstack [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1300.091565] env[65121]: WARNING neutronclient.v2_0.client [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1300.092530] env[65121]: WARNING openstack [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1300.093013] env[65121]: WARNING openstack [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1300.175403] env[65121]: DEBUG nova.network.neutron [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updating instance_info_cache with network_info: [{"id": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "address": "fa:16:3e:2b:98:34", "network": {"id": "7f1070a0-c22b-447f-a477-69aa08200805", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971658037-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72237a0d762645588c41231b0a34a796", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6966f473-59ac-49bb-9b7a-22c61f4e61e2", "external-id": "nsx-vlan-transportzone-668", "segmentation_id": 668, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d86d6-62", "ovs_interfaceid": "da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1300.383297] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1300.407321] env[65121]: INFO nova.scheduler.client.report [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Deleted allocations for instance 999ab257-c116-4a73-9b87-b52aeb9a3bb3 [ 1300.424974] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107654, 'name': Rename_Task, 'duration_secs': 0.143477} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.424974] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1300.424974] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b56632cc-d644-4b12-8c9a-acd4990c62c0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.430352] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1300.430352] env[65121]: value = "task-5107655" [ 1300.430352] env[65121]: _type = "Task" [ 1300.430352] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.438518] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.677927] env[65121]: DEBUG oslo_concurrency.lockutils [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Releasing lock "refresh_cache-b669fb5d-a68f-4390-b903-80879c239b4e" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1300.679144] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e12e28-23d0-403a-92aa-5ce825f68e29 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.686538] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Resuming the VM {{(pid=65121) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1300.686794] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85755050-44eb-4ff1-b3e3-51d8188ee096 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.693994] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1300.693994] env[65121]: value = "task-5107656" [ 1300.693994] env[65121]: _type = "Task" [ 1300.693994] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.702523] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107656, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.921835] env[65121]: DEBUG oslo_concurrency.lockutils [None req-f8f9d987-7e9f-434a-8da9-302039c8c8fc tempest-AttachVolumeShelveTestJSON-1482826837 tempest-AttachVolumeShelveTestJSON-1482826837-project-member] Lock "999ab257-c116-4a73-9b87-b52aeb9a3bb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.967s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1300.940740] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.205076] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107656, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.441477] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.626555] env[65121]: DEBUG oslo_concurrency.lockutils [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1301.627168] env[65121]: DEBUG oslo_concurrency.lockutils [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1301.705273] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107656, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.942802] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.131038] env[65121]: DEBUG nova.compute.utils [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1302.206308] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107656, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.443377] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.635983] env[65121]: DEBUG oslo_concurrency.lockutils [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1302.705582] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107656, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.944637] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.140285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "91a36110-5e97-4553-90c3-3e4213e30c14" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1303.140285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1303.140285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1303.140285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1303.140285] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1303.143012] env[65121]: INFO nova.compute.manager [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Terminating instance [ 1303.208471] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107656, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.445860] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.647868] env[65121]: DEBUG nova.compute.manager [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1303.648231] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1303.648614] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff10cea7-bf68-428a-b601-eaef0588cd9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.655863] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1303.655863] env[65121]: value = "task-5107658" [ 1303.655863] env[65121]: _type = "Task" [ 1303.655863] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.665306] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.696536] env[65121]: DEBUG oslo_concurrency.lockutils [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1303.696779] env[65121]: DEBUG oslo_concurrency.lockutils [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1303.697068] env[65121]: INFO nova.compute.manager [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Attaching volume 50c5e58e-27db-41c7-946f-1d44206ae552 to /dev/sdb [ 1303.709410] env[65121]: DEBUG oslo_vmware.api [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107656, 'name': PowerOnVM_Task, 'duration_secs': 2.753233} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.709830] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Resumed the VM {{(pid=65121) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1303.710119] env[65121]: DEBUG nova.compute.manager [None req-6fde5bce-5dac-4d55-b820-d25b0d0ed9fc tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1303.711331] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b8a321-dd9f-433d-803a-d7af47cb6b9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.737640] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58a9026-7004-4089-9b5a-071f25091650 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.746010] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb66df2-617f-4dab-99d1-e5fcc6e9651d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.762220] env[65121]: DEBUG nova.virt.block_device [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updating existing volume attachment record: f7d4128e-3d24-4efc-bccb-ab43ef65dc7e {{(pid=65121) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1303.945944] env[65121]: DEBUG oslo_vmware.api [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107655, 'name': PowerOnVM_Task, 'duration_secs': 3.019464} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.946991] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1304.053361] env[65121]: DEBUG nova.compute.manager [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1304.054519] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93fd70a-ae90-401e-b529-d61b4982dbb5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.170847] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107658, 'name': PowerOffVM_Task, 'duration_secs': 0.19335} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.171140] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1304.171333] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1304.171521] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993593', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'name': 'volume-47c28970-ebc3-443e-a72c-4ac58b364e40', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '91a36110-5e97-4553-90c3-3e4213e30c14', 'attached_at': '2025-12-12T14:31:25.000000', 'detached_at': '', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'serial': '47c28970-ebc3-443e-a72c-4ac58b364e40'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1304.172329] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a8097a-7f0e-458d-a025-96a6c1e11504 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.191782] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc74ca2-0fe5-4d9c-9675-08e204285110 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.199513] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675af84f-318c-4d35-8982-5d6370100d50 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.218565] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0422f6c7-3617-4f2e-acb7-54cf2370e35a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.236970] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] The volume has not been displaced from its original location: [datastore1] volume-47c28970-ebc3-443e-a72c-4ac58b364e40/volume-47c28970-ebc3-443e-a72c-4ac58b364e40.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1304.242454] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfiguring VM instance instance-0000007b to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1304.243833] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d462e6d-befb-4ed3-9425-08c45fe574c1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.270702] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1304.270702] env[65121]: value = "task-5107660" [ 1304.270702] env[65121]: _type = "Task" [ 1304.270702] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.283841] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107660, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.570722] env[65121]: DEBUG oslo_concurrency.lockutils [None req-0550287b-7b8e-423d-bcab-8ba665c380d2 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.583s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1304.780978] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107660, 'name': ReconfigVM_Task, 'duration_secs': 0.209674} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.781292] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Reconfigured VM instance instance-0000007b to detach disk 2000 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1304.785980] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6e36e82-6d3c-4f9b-850b-c59a1a8367cb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.801622] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1304.801622] env[65121]: value = "task-5107661" [ 1304.801622] env[65121]: _type = "Task" [ 1304.801622] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.810195] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.946636] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "b669fb5d-a68f-4390-b903-80879c239b4e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1304.946986] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1304.947114] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "b669fb5d-a68f-4390-b903-80879c239b4e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1304.947291] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1304.947454] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1304.949989] env[65121]: INFO nova.compute.manager [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Terminating instance [ 1305.312848] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107661, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.454237] env[65121]: DEBUG nova.compute.manager [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1305.454525] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1305.455462] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49cfab1-8a67-4ec1-bd91-e58a1f8a97f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.464456] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1305.464735] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94353720-0b70-412a-b272-5668fe600f09 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.472908] env[65121]: DEBUG oslo_vmware.api [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1305.472908] env[65121]: value = "task-5107662" [ 1305.472908] env[65121]: _type = "Task" [ 1305.472908] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.482939] env[65121]: DEBUG oslo_vmware.api [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107662, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.812705] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107661, 'name': ReconfigVM_Task, 'duration_secs': 0.766928} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.812970] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993593', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'name': 'volume-47c28970-ebc3-443e-a72c-4ac58b364e40', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '91a36110-5e97-4553-90c3-3e4213e30c14', 'attached_at': '2025-12-12T14:31:25.000000', 'detached_at': '', 'volume_id': '47c28970-ebc3-443e-a72c-4ac58b364e40', 'serial': '47c28970-ebc3-443e-a72c-4ac58b364e40'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1305.813272] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1305.814084] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb84a94b-e585-48d8-830d-14aa503688d4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.821703] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1305.821949] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23c86772-a623-4646-a14c-5b1c661c9399 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.840114] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1305.840338] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1305.840558] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1305.840735] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1305.840892] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1305.843194] env[65121]: INFO nova.compute.manager [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Terminating instance [ 1305.896878] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1305.897074] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1305.897258] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleting the datastore file [datastore1] 91a36110-5e97-4553-90c3-3e4213e30c14 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1305.897541] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5ada452-6878-4872-8f2b-6dda815b0c77 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.905880] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1305.905880] env[65121]: value = "task-5107664" [ 1305.905880] env[65121]: _type = "Task" [ 1305.905880] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.914640] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107664, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.982832] env[65121]: DEBUG oslo_vmware.api [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107662, 'name': PowerOffVM_Task, 'duration_secs': 0.197998} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.983135] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1305.983305] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1305.983563] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9113e6c2-e07c-4c10-809f-7370be4164b5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.057317] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1306.057548] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1306.057727] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleting the datastore file [datastore1] b669fb5d-a68f-4390-b903-80879c239b4e {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1306.058021] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-feac8293-4346-4f74-9c38-b45adb76a5ae {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.067030] env[65121]: DEBUG oslo_vmware.api [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for the task: (returnval){ [ 1306.067030] env[65121]: value = "task-5107667" [ 1306.067030] env[65121]: _type = "Task" [ 1306.067030] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.078645] env[65121]: DEBUG oslo_vmware.api [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.347488] env[65121]: DEBUG nova.compute.manager [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1306.347659] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1306.348626] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06fe22d-bb9d-4ce4-9335-30162ab76002 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.356844] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1306.357028] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca53d6b5-e4c3-49b9-a4ad-4234aee1c658 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.364363] env[65121]: DEBUG oslo_vmware.api [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1306.364363] env[65121]: value = "task-5107668" [ 1306.364363] env[65121]: _type = "Task" [ 1306.364363] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.374028] env[65121]: DEBUG oslo_vmware.api [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.416170] env[65121]: DEBUG oslo_vmware.api [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107664, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114513} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.416447] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1306.416598] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1306.416772] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1306.416948] env[65121]: INFO nova.compute.manager [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Took 2.77 seconds to destroy the instance on the hypervisor. [ 1306.417215] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1306.417410] env[65121]: DEBUG nova.compute.manager [-] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1306.417513] env[65121]: DEBUG nova.network.neutron [-] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1306.417745] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1306.418413] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.418669] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1306.487501] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1306.577627] env[65121]: DEBUG oslo_vmware.api [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Task: {'id': task-5107667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142688} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.577880] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1306.578060] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1306.578282] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1306.578460] env[65121]: INFO nova.compute.manager [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1306.578692] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1306.578881] env[65121]: DEBUG nova.compute.manager [-] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1306.578975] env[65121]: DEBUG nova.network.neutron [-] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1306.579220] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1306.579726] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.580012] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1306.875670] env[65121]: DEBUG oslo_vmware.api [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107668, 'name': PowerOffVM_Task, 'duration_secs': 0.190108} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.876219] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1306.876219] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1306.876398] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8510fdc3-defb-48d6-bae7-08b325615c67 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.944858] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1306.945149] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1306.945292] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleting the datastore file [datastore2] db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1306.945569] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fc00422-cb28-4c16-973a-d28b7eeb2d39 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.955490] env[65121]: DEBUG oslo_vmware.api [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for the task: (returnval){ [ 1306.955490] env[65121]: value = "task-5107670" [ 1306.955490] env[65121]: _type = "Task" [ 1306.955490] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.965281] env[65121]: DEBUG oslo_vmware.api [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.974190] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1307.080229] env[65121]: DEBUG nova.compute.manager [req-fe710b9f-6d40-4d29-9f99-352458b0a0d6 req-aa8b174e-20bc-4348-901a-4219c745785c service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Received event network-vif-deleted-76e75737-0e97-46ba-89f7-b94bbcbff126 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1307.080545] env[65121]: INFO nova.compute.manager [req-fe710b9f-6d40-4d29-9f99-352458b0a0d6 req-aa8b174e-20bc-4348-901a-4219c745785c service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Neutron deleted interface 76e75737-0e97-46ba-89f7-b94bbcbff126; detaching it from the instance and deleting it from the info cache [ 1307.080595] env[65121]: DEBUG nova.network.neutron [req-fe710b9f-6d40-4d29-9f99-352458b0a0d6 req-aa8b174e-20bc-4348-901a-4219c745785c service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1307.386466] env[65121]: DEBUG nova.network.neutron [-] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1307.473285] env[65121]: DEBUG oslo_vmware.api [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Task: {'id': task-5107670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149737} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.479017] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1307.479017] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1307.479017] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1307.479017] env[65121]: INFO nova.compute.manager [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1307.479017] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1307.479017] env[65121]: DEBUG nova.compute.manager [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1307.479017] env[65121]: DEBUG nova.network.neutron [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1307.479017] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1307.479017] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1307.479609] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.529672] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1307.585238] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8da0e5b7-03dc-4f7d-be01-921b1c137fdd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.601177] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70873b5-9bd0-4027-9c60-1267055b0ba0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.645039] env[65121]: DEBUG nova.compute.manager [req-fe710b9f-6d40-4d29-9f99-352458b0a0d6 req-aa8b174e-20bc-4348-901a-4219c745785c service nova] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Detach interface failed, port_id=76e75737-0e97-46ba-89f7-b94bbcbff126, reason: Instance 91a36110-5e97-4553-90c3-3e4213e30c14 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1307.893309] env[65121]: INFO nova.compute.manager [-] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Took 1.48 seconds to deallocate network for instance. [ 1307.987899] env[65121]: DEBUG nova.network.neutron [-] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1308.314625] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Volume attach. Driver type: vmdk {{(pid=65121) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1308.314883] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993612', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'name': 'volume-50c5e58e-27db-41c7-946f-1d44206ae552', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '07ff0535-f83c-4a1b-8077-26a8b1fa02b2', 'attached_at': '', 'detached_at': '', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'serial': '50c5e58e-27db-41c7-946f-1d44206ae552'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1308.315853] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7bb7cf-e639-405e-8da3-fe412fb66d70 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.334180] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629158c2-1762-4abd-a0d3-d1b380ff4df5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.360369] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] volume-50c5e58e-27db-41c7-946f-1d44206ae552/volume-50c5e58e-27db-41c7-946f-1d44206ae552.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1308.360627] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17b6ac87-23e1-47bb-a936-1b5ce6caabcb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.379892] env[65121]: DEBUG oslo_vmware.api [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1308.379892] env[65121]: value = "task-5107671" [ 1308.379892] env[65121]: _type = "Task" [ 1308.379892] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.392940] env[65121]: DEBUG oslo_vmware.api [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107671, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.440691] env[65121]: INFO nova.compute.manager [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Took 0.55 seconds to detach 1 volumes for instance. [ 1308.443609] env[65121]: DEBUG nova.compute.manager [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 91a36110-5e97-4553-90c3-3e4213e30c14] Deleting volume: 47c28970-ebc3-443e-a72c-4ac58b364e40 {{(pid=65121) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 1308.487723] env[65121]: DEBUG nova.network.neutron [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1308.491034] env[65121]: INFO nova.compute.manager [-] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Took 1.91 seconds to deallocate network for instance. [ 1308.890224] env[65121]: DEBUG oslo_vmware.api [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107671, 'name': ReconfigVM_Task, 'duration_secs': 0.422318} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.890501] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Reconfigured VM instance instance-0000007d to attach disk [datastore2] volume-50c5e58e-27db-41c7-946f-1d44206ae552/volume-50c5e58e-27db-41c7-946f-1d44206ae552.vmdk or device None with type thin {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1308.895233] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0d15ad3-0e6f-4980-81c7-38a938858270 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.911096] env[65121]: DEBUG oslo_vmware.api [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1308.911096] env[65121]: value = "task-5107673" [ 1308.911096] env[65121]: _type = "Task" [ 1308.911096] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.920589] env[65121]: DEBUG oslo_vmware.api [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.991863] env[65121]: INFO nova.compute.manager [-] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Took 1.51 seconds to deallocate network for instance. [ 1308.998259] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1308.998555] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1308.998765] env[65121]: DEBUG nova.objects.instance [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lazy-loading 'resources' on Instance uuid b669fb5d-a68f-4390-b903-80879c239b4e {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1309.010673] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1309.109774] env[65121]: DEBUG nova.compute.manager [req-3553b376-8516-413c-b6a8-29ec2008a176 req-99dc58db-939e-4246-8dc2-cc7e1dfecaad service nova] [instance: b669fb5d-a68f-4390-b903-80879c239b4e] Received event network-vif-deleted-da5d86d6-62e5-4fdb-b2a5-1ef06c2959a4 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1309.109900] env[65121]: DEBUG nova.compute.manager [req-3553b376-8516-413c-b6a8-29ec2008a176 req-99dc58db-939e-4246-8dc2-cc7e1dfecaad service nova] [instance: db417aa7-d170-45bc-b2c4-7a97400de1d8] Received event network-vif-deleted-3cbde80d-7df5-4785-821e-b45797716ab3 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1309.421054] env[65121]: DEBUG oslo_vmware.api [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107673, 'name': ReconfigVM_Task, 'duration_secs': 0.180829} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.421506] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993612', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'name': 'volume-50c5e58e-27db-41c7-946f-1d44206ae552', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '07ff0535-f83c-4a1b-8077-26a8b1fa02b2', 'attached_at': '', 'detached_at': '', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'serial': '50c5e58e-27db-41c7-946f-1d44206ae552'} {{(pid=65121) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1309.499839] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1309.615792] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fe4f7f-e629-4030-8234-4faf13e231c4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.623479] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7950ec5f-94e4-4cc8-81af-ad8e2630fd0b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.655763] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ece184-4c89-4dc7-8db8-da796e12c17a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.665050] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd45538-efb1-4984-ba3d-5fc0049237df {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.679063] env[65121]: DEBUG nova.compute.provider_tree [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.182528] env[65121]: DEBUG nova.scheduler.client.report [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1310.458184] env[65121]: DEBUG nova.objects.instance [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'flavor' on Instance uuid 07ff0535-f83c-4a1b-8077-26a8b1fa02b2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1310.689301] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.691s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1310.692374] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.681s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1310.692615] env[65121]: DEBUG nova.objects.instance [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'resources' on Instance uuid 91a36110-5e97-4553-90c3-3e4213e30c14 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1310.710371] env[65121]: INFO nova.scheduler.client.report [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Deleted allocations for instance b669fb5d-a68f-4390-b903-80879c239b4e [ 1310.963372] env[65121]: DEBUG oslo_concurrency.lockutils [None req-374a86d1-6a2d-4e91-ae07-098e193ce413 tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.267s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1311.176811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1311.177175] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1311.217513] env[65121]: DEBUG oslo_concurrency.lockutils [None req-77bf5c04-32f6-49fb-bcd4-9cb945f65c57 tempest-ServerActionsTestJSON-460922022 tempest-ServerActionsTestJSON-460922022-project-member] Lock "b669fb5d-a68f-4390-b903-80879c239b4e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.270s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1311.310557] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f35c90-b2e2-44a4-afff-18985d317c75 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.318465] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa743d68-7534-419f-9d97-8029dc96abfc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.351363] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9a06a9-b862-4bb8-ab63-1d145a85ab31 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.359209] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7136ced-17d8-41a5-8732-b04195d77acb {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.373353] env[65121]: DEBUG nova.compute.provider_tree [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1311.681125] env[65121]: INFO nova.compute.manager [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Detaching volume 50c5e58e-27db-41c7-946f-1d44206ae552 [ 1311.717010] env[65121]: INFO nova.virt.block_device [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Attempting to driver detach volume 50c5e58e-27db-41c7-946f-1d44206ae552 from mountpoint /dev/sdb [ 1311.717336] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1311.717605] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993612', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'name': 'volume-50c5e58e-27db-41c7-946f-1d44206ae552', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '07ff0535-f83c-4a1b-8077-26a8b1fa02b2', 'attached_at': '', 'detached_at': '', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'serial': '50c5e58e-27db-41c7-946f-1d44206ae552'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1311.718763] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2272a4-d9a4-4084-be4c-787dc1cb26f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.742783] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14293687-0ecb-4526-a590-e8a4f68e2cb5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.750884] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8f3c85-010f-479c-bf06-cf14b5601297 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.773712] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547d0380-9c9f-484c-afde-e963507acc3e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.790024] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] The volume has not been displaced from its original location: [datastore2] volume-50c5e58e-27db-41c7-946f-1d44206ae552/volume-50c5e58e-27db-41c7-946f-1d44206ae552.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1311.795820] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Reconfiguring VM instance instance-0000007d to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1311.796223] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b1b75f7-5b27-4e85-aa09-9275679c8889 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.816067] env[65121]: DEBUG oslo_vmware.api [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1311.816067] env[65121]: value = "task-5107674" [ 1311.816067] env[65121]: _type = "Task" [ 1311.816067] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.824596] env[65121]: DEBUG oslo_vmware.api [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107674, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.877023] env[65121]: DEBUG nova.scheduler.client.report [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1312.328440] env[65121]: DEBUG oslo_vmware.api [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107674, 'name': ReconfigVM_Task, 'duration_secs': 0.240412} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.328440] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Reconfigured VM instance instance-0000007d to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1312.332905] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d65a6972-54a3-4f64-b8ab-a6ebfb39800f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.348447] env[65121]: DEBUG oslo_vmware.api [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1312.348447] env[65121]: value = "task-5107675" [ 1312.348447] env[65121]: _type = "Task" [ 1312.348447] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.361778] env[65121]: DEBUG oslo_vmware.api [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107675, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.383288] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.691s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1312.386776] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.886s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1312.386776] env[65121]: DEBUG nova.objects.instance [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lazy-loading 'resources' on Instance uuid db417aa7-d170-45bc-b2c4-7a97400de1d8 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1312.416631] env[65121]: INFO nova.scheduler.client.report [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted allocations for instance 91a36110-5e97-4553-90c3-3e4213e30c14 [ 1312.858758] env[65121]: DEBUG oslo_vmware.api [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107675, 'name': ReconfigVM_Task, 'duration_secs': 0.168996} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.859200] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993612', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'name': 'volume-50c5e58e-27db-41c7-946f-1d44206ae552', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '07ff0535-f83c-4a1b-8077-26a8b1fa02b2', 'attached_at': '', 'detached_at': '', 'volume_id': '50c5e58e-27db-41c7-946f-1d44206ae552', 'serial': '50c5e58e-27db-41c7-946f-1d44206ae552'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1312.930735] env[65121]: DEBUG oslo_concurrency.lockutils [None req-5cbde55b-f7cc-4e3f-943b-3fc4c9aa0ba8 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "91a36110-5e97-4553-90c3-3e4213e30c14" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.790s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1313.006521] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892a082c-158a-43f4-9072-943f0b57fb36 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.015630] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4101a4-af98-404a-9222-c8e6d14b6375 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.049836] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d54f492-2c97-4137-811d-52a554d21364 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.059297] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cd44df-552b-4bec-8a71-b7333f6c866d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.075857] env[65121]: DEBUG nova.compute.provider_tree [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1313.290612] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "992ba328-206a-4d47-a28b-f3ef9b20817f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.290819] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1313.291066] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "992ba328-206a-4d47-a28b-f3ef9b20817f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.291260] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1313.291449] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1313.294069] env[65121]: INFO nova.compute.manager [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Terminating instance [ 1313.404221] env[65121]: DEBUG nova.objects.instance [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'flavor' on Instance uuid 07ff0535-f83c-4a1b-8077-26a8b1fa02b2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1313.578931] env[65121]: DEBUG nova.scheduler.client.report [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1313.798106] env[65121]: DEBUG nova.compute.manager [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1313.798462] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1313.799398] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79242a00-a12c-4806-b749-f8b0db9f7c76 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.808698] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1313.809059] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a35047d5-65e0-4279-8edf-0c26ba526d46 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.817030] env[65121]: DEBUG oslo_vmware.api [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1313.817030] env[65121]: value = "task-5107676" [ 1313.817030] env[65121]: _type = "Task" [ 1313.817030] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.827465] env[65121]: DEBUG oslo_vmware.api [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.084613] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1314.108453] env[65121]: INFO nova.scheduler.client.report [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Deleted allocations for instance db417aa7-d170-45bc-b2c4-7a97400de1d8 [ 1314.327895] env[65121]: DEBUG oslo_vmware.api [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107676, 'name': PowerOffVM_Task, 'duration_secs': 0.226709} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.328206] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1314.328359] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1314.328618] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0323a5a5-d58b-431c-a9bd-5451a0bc0139 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.403408] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1314.403721] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1314.404026] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleting the datastore file [datastore1] 992ba328-206a-4d47-a28b-f3ef9b20817f {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1314.404478] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e730d8a-bf60-4c13-8996-7467948262f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.412704] env[65121]: DEBUG oslo_concurrency.lockutils [None req-e4413fbe-9436-466a-bc50-a2fd26e441dd tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.235s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1314.415510] env[65121]: DEBUG oslo_vmware.api [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1314.415510] env[65121]: value = "task-5107678" [ 1314.415510] env[65121]: _type = "Task" [ 1314.415510] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.425828] env[65121]: DEBUG oslo_vmware.api [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.617833] env[65121]: DEBUG oslo_concurrency.lockutils [None req-aec8e632-c03f-450e-88c2-4b99984f6be4 tempest-ServerActionsTestOtherB-1773662609 tempest-ServerActionsTestOtherB-1773662609-project-member] Lock "db417aa7-d170-45bc-b2c4-7a97400de1d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.777s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1314.927561] env[65121]: DEBUG oslo_vmware.api [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139453} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.928027] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1314.928027] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1314.928250] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1314.928411] env[65121]: INFO nova.compute.manager [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1314.928653] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1314.928841] env[65121]: DEBUG nova.compute.manager [-] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1314.928945] env[65121]: DEBUG nova.network.neutron [-] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1314.929241] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1314.929710] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1314.929963] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1314.980741] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1315.304531] env[65121]: DEBUG nova.compute.manager [req-8118de92-f131-488f-b8b4-d7c3fe918268 req-115a4736-b7e7-4abd-aac2-4f7f22fbd28a service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Received event network-vif-deleted-747792c4-6fe8-4e3c-8829-6060b4a047f0 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1315.304823] env[65121]: INFO nova.compute.manager [req-8118de92-f131-488f-b8b4-d7c3fe918268 req-115a4736-b7e7-4abd-aac2-4f7f22fbd28a service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Neutron deleted interface 747792c4-6fe8-4e3c-8829-6060b4a047f0; detaching it from the instance and deleting it from the info cache [ 1315.305024] env[65121]: DEBUG nova.network.neutron [req-8118de92-f131-488f-b8b4-d7c3fe918268 req-115a4736-b7e7-4abd-aac2-4f7f22fbd28a service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1315.511968] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1315.512256] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1315.512461] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1315.512660] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1315.512837] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1315.515107] env[65121]: INFO nova.compute.manager [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Terminating instance [ 1315.698628] env[65121]: DEBUG nova.network.neutron [-] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1315.810114] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18dad6ad-3dfd-4d73-8708-b2ab69ff3bcf {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.820113] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ef9af8-771e-4090-bcba-58c62fda2c98 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.853325] env[65121]: DEBUG nova.compute.manager [req-8118de92-f131-488f-b8b4-d7c3fe918268 req-115a4736-b7e7-4abd-aac2-4f7f22fbd28a service nova] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Detach interface failed, port_id=747792c4-6fe8-4e3c-8829-6060b4a047f0, reason: Instance 992ba328-206a-4d47-a28b-f3ef9b20817f could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1316.021353] env[65121]: DEBUG nova.compute.manager [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1316.021618] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1316.022502] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19a9955-8e4f-4dc6-9e64-e85c01f2d8ee {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.030870] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1316.031193] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-378033f3-2ea5-451f-b457-7d15b7f61f4f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.040190] env[65121]: DEBUG oslo_vmware.api [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1316.040190] env[65121]: value = "task-5107680" [ 1316.040190] env[65121]: _type = "Task" [ 1316.040190] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.049554] env[65121]: DEBUG oslo_vmware.api [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107680, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.201436] env[65121]: INFO nova.compute.manager [-] [instance: 992ba328-206a-4d47-a28b-f3ef9b20817f] Took 1.27 seconds to deallocate network for instance. [ 1316.551522] env[65121]: DEBUG oslo_vmware.api [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107680, 'name': PowerOffVM_Task, 'duration_secs': 0.228892} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.551846] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1316.551954] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1316.552237] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8bcd8394-bd01-46fe-bf9e-a2ec549c3139 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.626098] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1316.626382] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1316.626609] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleting the datastore file [datastore1] 07ff0535-f83c-4a1b-8077-26a8b1fa02b2 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1316.626926] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64e82f24-880a-4e93-8e6c-1326ed23aa1b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.635747] env[65121]: DEBUG oslo_vmware.api [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for the task: (returnval){ [ 1316.635747] env[65121]: value = "task-5107682" [ 1316.635747] env[65121]: _type = "Task" [ 1316.635747] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.646925] env[65121]: DEBUG oslo_vmware.api [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107682, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.711828] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1316.712141] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1316.712383] env[65121]: DEBUG nova.objects.instance [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'resources' on Instance uuid 992ba328-206a-4d47-a28b-f3ef9b20817f {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1317.149296] env[65121]: DEBUG oslo_vmware.api [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Task: {'id': task-5107682, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139479} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.149599] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1317.149784] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1317.149953] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1317.150136] env[65121]: INFO nova.compute.manager [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1317.150432] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1317.150682] env[65121]: DEBUG nova.compute.manager [-] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1317.150722] env[65121]: DEBUG nova.network.neutron [-] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1317.150962] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1317.151530] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1317.151788] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1317.190877] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1317.339034] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a826d69e-bc03-484f-b7b1-4224c8506272 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.347651] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df39a7bb-ed70-4898-90c6-703cb3d13c21 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.383102] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54eaa727-0538-4807-9020-177191bc3615 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.392474] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc04623-7eb2-4b7a-9784-d961d9de4006 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.405303] env[65121]: DEBUG nova.compute.provider_tree [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1317.756851] env[65121]: DEBUG nova.compute.manager [req-c6a1e773-ffb5-4ce3-ba9c-914b8d195df0 req-02b1be30-7436-4df5-a477-cbbae8ec2d59 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Received event network-vif-deleted-1603a551-b060-4f2e-a724-cfcd68db842a {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1317.760088] env[65121]: INFO nova.compute.manager [req-c6a1e773-ffb5-4ce3-ba9c-914b8d195df0 req-02b1be30-7436-4df5-a477-cbbae8ec2d59 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Neutron deleted interface 1603a551-b060-4f2e-a724-cfcd68db842a; detaching it from the instance and deleting it from the info cache [ 1317.761805] env[65121]: DEBUG nova.network.neutron [req-c6a1e773-ffb5-4ce3-ba9c-914b8d195df0 req-02b1be30-7436-4df5-a477-cbbae8ec2d59 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1317.911035] env[65121]: DEBUG nova.scheduler.client.report [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1318.201236] env[65121]: DEBUG nova.network.neutron [-] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1318.266102] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f69436f-3794-4b9c-93c8-514ad4a936d2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.277759] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579a3ffa-1d91-40f5-8267-511965eb6bf1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.309526] env[65121]: DEBUG nova.compute.manager [req-c6a1e773-ffb5-4ce3-ba9c-914b8d195df0 req-02b1be30-7436-4df5-a477-cbbae8ec2d59 service nova] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Detach interface failed, port_id=1603a551-b060-4f2e-a724-cfcd68db842a, reason: Instance 07ff0535-f83c-4a1b-8077-26a8b1fa02b2 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1318.415076] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.703s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1318.442273] env[65121]: INFO nova.scheduler.client.report [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted allocations for instance 992ba328-206a-4d47-a28b-f3ef9b20817f [ 1318.704805] env[65121]: INFO nova.compute.manager [-] [instance: 07ff0535-f83c-4a1b-8077-26a8b1fa02b2] Took 1.55 seconds to deallocate network for instance. [ 1318.953953] env[65121]: DEBUG oslo_concurrency.lockutils [None req-328f709a-e9ec-4066-96c4-7256b919fbaf tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "992ba328-206a-4d47-a28b-f3ef9b20817f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.663s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1319.217391] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.217391] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.217391] env[65121]: DEBUG nova.objects.instance [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lazy-loading 'resources' on Instance uuid 07ff0535-f83c-4a1b-8077-26a8b1fa02b2 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1319.847726] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ec2fa4-8185-4bf0-9d14-f5d2b8adcc2d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.857086] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb94e78-8965-4905-bd26-adca3cb17c14 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.893260] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf73dcc-fbb3-4207-baf3-46f43db5823a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.902320] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29915941-e97d-44cc-ab86-be74fe69bfcc {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.919139] env[65121]: DEBUG nova.compute.provider_tree [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1319.999015] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.999323] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1320.422042] env[65121]: DEBUG nova.scheduler.client.report [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1320.505081] env[65121]: DEBUG nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Starting instance... {{(pid=65121) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1320.927879] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1320.950664] env[65121]: INFO nova.scheduler.client.report [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Deleted allocations for instance 07ff0535-f83c-4a1b-8077-26a8b1fa02b2 [ 1321.027146] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1321.027450] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1321.029009] env[65121]: INFO nova.compute.claims [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1321.459090] env[65121]: DEBUG oslo_concurrency.lockutils [None req-9886c055-7e3f-417d-a466-244e4dd990cf tempest-AttachVolumeNegativeTest-1790674778 tempest-AttachVolumeNegativeTest-1790674778-project-member] Lock "07ff0535-f83c-4a1b-8077-26a8b1fa02b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.947s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1321.485513] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1321.485742] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1321.485935] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "4a72fedd-b114-468e-8f34-0caec6ce73fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1321.486154] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1321.486287] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1321.488294] env[65121]: INFO nova.compute.manager [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Terminating instance [ 1321.992426] env[65121]: DEBUG nova.compute.manager [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1321.992666] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1321.993598] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfc4ebd-86af-4133-aa9c-637c0c8e7c12 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.002040] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1322.002298] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a71385a-8777-4d32-ac47-076f55247557 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.009486] env[65121]: DEBUG oslo_vmware.api [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1322.009486] env[65121]: value = "task-5107684" [ 1322.009486] env[65121]: _type = "Task" [ 1322.009486] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.019342] env[65121]: DEBUG oslo_vmware.api [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107684, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.125464] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46e5954-da43-4618-a510-a8712f1cfea9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.133777] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dacb87f8-0686-439c-95e2-3c4961f9bd78 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.171830] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf7d42b-418c-4165-a929-13ec50bfd319 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.184351] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7f21a0-9d36-4d50-9197-b2bf2b4584e9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.200561] env[65121]: DEBUG nova.compute.provider_tree [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1322.520030] env[65121]: DEBUG oslo_vmware.api [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107684, 'name': PowerOffVM_Task, 'duration_secs': 0.214678} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.520196] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1322.520313] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1322.520560] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0afbdb3-1b37-4989-a319-594c6027b37f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.590449] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1322.590870] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1322.590870] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleting the datastore file [datastore2] 4a72fedd-b114-468e-8f34-0caec6ce73fa {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1322.591746] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd0af44c-c498-4d33-b9a6-3fd67bfa777b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.599762] env[65121]: DEBUG oslo_vmware.api [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for the task: (returnval){ [ 1322.599762] env[65121]: value = "task-5107687" [ 1322.599762] env[65121]: _type = "Task" [ 1322.599762] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.608318] env[65121]: DEBUG oslo_vmware.api [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107687, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.703772] env[65121]: DEBUG nova.scheduler.client.report [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1323.109734] env[65121]: DEBUG oslo_vmware.api [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Task: {'id': task-5107687, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131955} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.110089] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1323.110298] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1323.110472] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1323.110647] env[65121]: INFO nova.compute.manager [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1323.110883] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1323.111095] env[65121]: DEBUG nova.compute.manager [-] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1323.111189] env[65121]: DEBUG nova.network.neutron [-] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1323.111432] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1323.111956] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1323.112217] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1323.194761] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1323.209860] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.182s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1323.210454] env[65121]: DEBUG nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Start building networks asynchronously for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1323.419816] env[65121]: DEBUG nova.compute.manager [req-d56adac5-eec5-4489-aab1-af9333c15ee0 req-bd244a74-3719-487a-916b-03e9c68087c7 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Received event network-vif-deleted-4304bec2-f452-4be6-aeda-3ad8354a4547 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1323.419816] env[65121]: INFO nova.compute.manager [req-d56adac5-eec5-4489-aab1-af9333c15ee0 req-bd244a74-3719-487a-916b-03e9c68087c7 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Neutron deleted interface 4304bec2-f452-4be6-aeda-3ad8354a4547; detaching it from the instance and deleting it from the info cache [ 1323.419961] env[65121]: DEBUG nova.network.neutron [req-d56adac5-eec5-4489-aab1-af9333c15ee0 req-bd244a74-3719-487a-916b-03e9c68087c7 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1323.715806] env[65121]: DEBUG nova.compute.utils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Using /dev/sd instead of None {{(pid=65121) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1323.717341] env[65121]: DEBUG nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Allocating IP information in the background. {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1323.717537] env[65121]: DEBUG nova.network.neutron [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] allocate_for_instance() {{(pid=65121) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1323.717854] env[65121]: WARNING neutronclient.v2_0.client [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1323.718174] env[65121]: WARNING neutronclient.v2_0.client [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1323.718782] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1323.719173] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1323.772407] env[65121]: DEBUG nova.policy [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67208ebf0be24f33b01d880879184210', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb97329518dd465f8b283d53e4aca0f9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65121) authorize /opt/stack/nova/nova/policy.py:192}} [ 1323.894248] env[65121]: DEBUG nova.network.neutron [-] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1323.922670] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b400648c-e2e3-488e-aaa7-cd74b26bb26d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.933098] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56626ab-82bd-4f5c-b294-80fc3366a7be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.967505] env[65121]: DEBUG nova.compute.manager [req-d56adac5-eec5-4489-aab1-af9333c15ee0 req-bd244a74-3719-487a-916b-03e9c68087c7 service nova] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Detach interface failed, port_id=4304bec2-f452-4be6-aeda-3ad8354a4547, reason: Instance 4a72fedd-b114-468e-8f34-0caec6ce73fa could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1324.031613] env[65121]: DEBUG nova.network.neutron [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Successfully created port: 1f005be0-5cc6-421b-8b7f-769561de8d43 {{(pid=65121) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1324.227263] env[65121]: DEBUG nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Start building block device mappings for instance. {{(pid=65121) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1324.397403] env[65121]: INFO nova.compute.manager [-] [instance: 4a72fedd-b114-468e-8f34-0caec6ce73fa] Took 1.29 seconds to deallocate network for instance. [ 1324.674500] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1324.674772] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1324.908380] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1324.908811] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1324.909179] env[65121]: DEBUG nova.objects.instance [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lazy-loading 'resources' on Instance uuid 4a72fedd-b114-468e-8f34-0caec6ce73fa {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1325.178381] env[65121]: INFO nova.compute.manager [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Detaching volume 50802f92-48fc-45a4-8fdb-0811fade10a3 [ 1325.214044] env[65121]: INFO nova.virt.block_device [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Attempting to driver detach volume 50802f92-48fc-45a4-8fdb-0811fade10a3 from mountpoint /dev/sdb [ 1325.214044] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Volume detach. Driver type: vmdk {{(pid=65121) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1325.214222] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993606', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'name': 'volume-50802f92-48fc-45a4-8fdb-0811fade10a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd34b8457-adc1-4d6b-befd-f6f1aff6d1fb', 'attached_at': '', 'detached_at': '', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'serial': '50802f92-48fc-45a4-8fdb-0811fade10a3'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1325.215139] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab0958d-0419-4a12-85bd-01ab90c96e8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.237812] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af062c37-3823-437c-841a-f4056aafc49e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.241650] env[65121]: DEBUG nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Start spawning the instance on the hypervisor. {{(pid=65121) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1325.248978] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f74c34f-bd4f-4827-afde-9747eb08e3f1 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.272599] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b94e9f-91cc-4d3d-9afd-d15bbfa0415b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.288575] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] The volume has not been displaced from its original location: [datastore1] volume-50802f92-48fc-45a4-8fdb-0811fade10a3/volume-50802f92-48fc-45a4-8fdb-0811fade10a3.vmdk. No consolidation needed. {{(pid=65121) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1325.294048] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfiguring VM instance instance-0000007a to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1325.296536] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T14:18:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T14:17:55Z,direct_url=,disk_format='vmdk',id=3ed5d656-a616-4609-8bb7-b02beb3ac3df,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='009aecc07c7c45ccbe2e2a9f13e8ce6b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T14:17:55Z,virtual_size=,visibility=), allow threads: False {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1325.296762] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Flavor limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1325.296907] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Image limits 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1325.297096] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Flavor pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1325.297237] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Image pref 0:0:0 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1325.297377] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65121) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1325.297660] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1325.297827] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1325.298054] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Got 1 possible topologies {{(pid=65121) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1325.298214] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1325.298423] env[65121]: DEBUG nova.virt.hardware [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65121) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1325.298735] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dbcddc1-36b0-42b9-a34e-daa15e83ebc6 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.312682] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34bd668-fb95-4f66-9ad5-bc2fbea3c9f4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.323081] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a611f505-75cf-49ba-97ba-a4a33b601d1d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.327081] env[65121]: DEBUG oslo_vmware.api [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1325.327081] env[65121]: value = "task-5107688" [ 1325.327081] env[65121]: _type = "Task" [ 1325.327081] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.346514] env[65121]: DEBUG oslo_vmware.api [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107688, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.484023] env[65121]: DEBUG nova.compute.manager [req-4ad1fc5f-1ef3-47be-b38f-31d9ce3b42f0 req-4fa9e67b-4e4e-42a2-aafa-c2e58c807cdf service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Received event network-vif-plugged-1f005be0-5cc6-421b-8b7f-769561de8d43 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1325.484290] env[65121]: DEBUG oslo_concurrency.lockutils [req-4ad1fc5f-1ef3-47be-b38f-31d9ce3b42f0 req-4fa9e67b-4e4e-42a2-aafa-c2e58c807cdf service nova] Acquiring lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1325.484478] env[65121]: DEBUG oslo_concurrency.lockutils [req-4ad1fc5f-1ef3-47be-b38f-31d9ce3b42f0 req-4fa9e67b-4e4e-42a2-aafa-c2e58c807cdf service nova] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1325.484779] env[65121]: DEBUG oslo_concurrency.lockutils [req-4ad1fc5f-1ef3-47be-b38f-31d9ce3b42f0 req-4fa9e67b-4e4e-42a2-aafa-c2e58c807cdf service nova] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1325.484909] env[65121]: DEBUG nova.compute.manager [req-4ad1fc5f-1ef3-47be-b38f-31d9ce3b42f0 req-4fa9e67b-4e4e-42a2-aafa-c2e58c807cdf service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] No waiting events found dispatching network-vif-plugged-1f005be0-5cc6-421b-8b7f-769561de8d43 {{(pid=65121) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1325.485050] env[65121]: WARNING nova.compute.manager [req-4ad1fc5f-1ef3-47be-b38f-31d9ce3b42f0 req-4fa9e67b-4e4e-42a2-aafa-c2e58c807cdf service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Received unexpected event network-vif-plugged-1f005be0-5cc6-421b-8b7f-769561de8d43 for instance with vm_state building and task_state spawning. [ 1325.503035] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79a4229-f452-4ba5-8776-faf30d20afb9 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.512239] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59dcd2e0-afe4-449a-9eca-ff15cd84237f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.546796] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b99f80d-d1b8-42bb-b856-15646a17b02b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.555156] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2291bf6d-2dc9-4232-b638-f5c3e904b74c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.570410] env[65121]: DEBUG nova.compute.provider_tree [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1325.595492] env[65121]: DEBUG nova.network.neutron [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Successfully updated port: 1f005be0-5cc6-421b-8b7f-769561de8d43 {{(pid=65121) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1325.840452] env[65121]: DEBUG oslo_vmware.api [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107688, 'name': ReconfigVM_Task, 'duration_secs': 0.266645} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.840744] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Reconfigured VM instance instance-0000007a to detach disk 2001 {{(pid=65121) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1325.847570] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-194b69bb-61e6-477f-88ec-02bc63745de0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.864694] env[65121]: DEBUG oslo_vmware.api [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1325.864694] env[65121]: value = "task-5107690" [ 1325.864694] env[65121]: _type = "Task" [ 1325.864694] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.875423] env[65121]: DEBUG oslo_vmware.api [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107690, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.075069] env[65121]: DEBUG nova.scheduler.client.report [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1326.096706] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "refresh_cache-55d1d720-e0ba-4047-9820-43e6f4bbbc85" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.096881] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquired lock "refresh_cache-55d1d720-e0ba-4047-9820-43e6f4bbbc85" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1326.096993] env[65121]: DEBUG nova.network.neutron [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Building network info cache for instance {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1326.375103] env[65121]: DEBUG oslo_vmware.api [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107690, 'name': ReconfigVM_Task, 'duration_secs': 0.149499} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.375461] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-993606', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'name': 'volume-50802f92-48fc-45a4-8fdb-0811fade10a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd34b8457-adc1-4d6b-befd-f6f1aff6d1fb', 'attached_at': '', 'detached_at': '', 'volume_id': '50802f92-48fc-45a4-8fdb-0811fade10a3', 'serial': '50802f92-48fc-45a4-8fdb-0811fade10a3'} {{(pid=65121) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1326.579299] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.670s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1326.599634] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1326.600199] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1326.610348] env[65121]: INFO nova.scheduler.client.report [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Deleted allocations for instance 4a72fedd-b114-468e-8f34-0caec6ce73fa [ 1326.648086] env[65121]: DEBUG nova.network.neutron [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Instance cache missing network info. {{(pid=65121) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3424}} [ 1326.670031] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1326.670433] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1326.740886] env[65121]: WARNING neutronclient.v2_0.client [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1326.741835] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1326.742129] env[65121]: WARNING openstack [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1326.829548] env[65121]: DEBUG nova.network.neutron [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Updating instance_info_cache with network_info: [{"id": "1f005be0-5cc6-421b-8b7f-769561de8d43", "address": "fa:16:3e:15:0d:eb", "network": {"id": "a7ab21c4-6c64-4567-a7c5-8dc8b4bcfa42", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-491088417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb97329518dd465f8b283d53e4aca0f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a75bb6e-6331-4429-b1b9-c968cc22b9c9", "external-id": "nsx-vlan-transportzone-244", "segmentation_id": 244, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f005be0-5c", "ovs_interfaceid": "1f005be0-5cc6-421b-8b7f-769561de8d43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1326.920108] env[65121]: DEBUG nova.objects.instance [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lazy-loading 'flavor' on Instance uuid d34b8457-adc1-4d6b-befd-f6f1aff6d1fb {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1327.116915] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3beef78c-1083-47c9-b4b6-df4fb8813ed4 tempest-ServerActionsTestOtherA-531735361 tempest-ServerActionsTestOtherA-531735361-project-member] Lock "4a72fedd-b114-468e-8f34-0caec6ce73fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.631s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1327.332439] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Releasing lock "refresh_cache-55d1d720-e0ba-4047-9820-43e6f4bbbc85" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1327.332850] env[65121]: DEBUG nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Instance network_info: |[{"id": "1f005be0-5cc6-421b-8b7f-769561de8d43", "address": "fa:16:3e:15:0d:eb", "network": {"id": "a7ab21c4-6c64-4567-a7c5-8dc8b4bcfa42", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-491088417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb97329518dd465f8b283d53e4aca0f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a75bb6e-6331-4429-b1b9-c968cc22b9c9", "external-id": "nsx-vlan-transportzone-244", "segmentation_id": 244, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f005be0-5c", "ovs_interfaceid": "1f005be0-5cc6-421b-8b7f-769561de8d43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65121) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1327.333365] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:0d:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a75bb6e-6331-4429-b1b9-c968cc22b9c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f005be0-5cc6-421b-8b7f-769561de8d43', 'vif_model': 'vmxnet3'}] {{(pid=65121) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1327.342215] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Creating folder: Project (fb97329518dd465f8b283d53e4aca0f9). Parent ref: group-v993268. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1327.342369] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-faccce46-5d32-4e3a-94dd-c65d7dab6204 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.355648] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Created folder: Project (fb97329518dd465f8b283d53e4aca0f9) in parent group-v993268. [ 1327.355865] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Creating folder: Instances. Parent ref: group-v993613. {{(pid=65121) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1327.356197] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-730f7215-d9ec-4681-92f7-b99fe92d1197 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.367144] env[65121]: INFO nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Created folder: Instances in parent group-v993613. [ 1327.367353] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1327.367556] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Creating VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1327.367771] env[65121]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6c39ce5d-3f33-4285-b71b-d9cff458c389 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.388948] env[65121]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1327.388948] env[65121]: value = "task-5107693" [ 1327.388948] env[65121]: _type = "Task" [ 1327.388948] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.397977] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107693, 'name': CreateVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.517445] env[65121]: DEBUG nova.compute.manager [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Received event network-changed-1f005be0-5cc6-421b-8b7f-769561de8d43 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1327.517690] env[65121]: DEBUG nova.compute.manager [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Refreshing instance network info cache due to event network-changed-1f005be0-5cc6-421b-8b7f-769561de8d43. {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1327.517906] env[65121]: DEBUG oslo_concurrency.lockutils [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Acquiring lock "refresh_cache-55d1d720-e0ba-4047-9820-43e6f4bbbc85" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1327.518096] env[65121]: DEBUG oslo_concurrency.lockutils [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Acquired lock "refresh_cache-55d1d720-e0ba-4047-9820-43e6f4bbbc85" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1327.518278] env[65121]: DEBUG nova.network.neutron [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Refreshing network info cache for port 1f005be0-5cc6-421b-8b7f-769561de8d43 {{(pid=65121) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1327.899745] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107693, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.928713] env[65121]: DEBUG oslo_concurrency.lockutils [None req-ee0a521e-cdec-43f3-a689-f0d72d1aa23f tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.254s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1328.021301] env[65121]: WARNING neutronclient.v2_0.client [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.022013] env[65121]: WARNING openstack [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.022394] env[65121]: WARNING openstack [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.134679] env[65121]: WARNING openstack [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.135092] env[65121]: WARNING openstack [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.195769] env[65121]: WARNING neutronclient.v2_0.client [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.196437] env[65121]: WARNING openstack [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.196787] env[65121]: WARNING openstack [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.275758] env[65121]: DEBUG nova.network.neutron [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Updated VIF entry in instance network info cache for port 1f005be0-5cc6-421b-8b7f-769561de8d43. {{(pid=65121) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3584}} [ 1328.276143] env[65121]: DEBUG nova.network.neutron [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Updating instance_info_cache with network_info: [{"id": "1f005be0-5cc6-421b-8b7f-769561de8d43", "address": "fa:16:3e:15:0d:eb", "network": {"id": "a7ab21c4-6c64-4567-a7c5-8dc8b4bcfa42", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-491088417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb97329518dd465f8b283d53e4aca0f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a75bb6e-6331-4429-b1b9-c968cc22b9c9", "external-id": "nsx-vlan-transportzone-244", "segmentation_id": 244, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f005be0-5c", "ovs_interfaceid": "1f005be0-5cc6-421b-8b7f-769561de8d43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1328.401616] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107693, 'name': CreateVM_Task} progress is 99%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.779304] env[65121]: DEBUG oslo_concurrency.lockutils [req-cdde74a1-4257-4189-ae81-1516d038bdaf req-e549dc4f-d4b3-4af5-91b0-3fea58dd423c service nova] Releasing lock "refresh_cache-55d1d720-e0ba-4047-9820-43e6f4bbbc85" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1328.902829] env[65121]: DEBUG oslo_vmware.api [-] Task: {'id': task-5107693, 'name': CreateVM_Task, 'duration_secs': 1.358727} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.903313] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Created VM on the ESX host {{(pid=65121) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1328.903883] env[65121]: WARNING neutronclient.v2_0.client [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.904106] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.904297] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1328.904653] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1328.905074] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d461ef5-94cb-49bc-9d84-4a88a2622dca {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.910490] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1328.910490] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]522e5429-c27e-525a-c617-251fb01fac13" [ 1328.910490] env[65121]: _type = "Task" [ 1328.910490] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.921614] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522e5429-c27e-525a-c617-251fb01fac13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.080622] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1329.080893] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1329.081163] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1329.081504] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1329.081738] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1329.084141] env[65121]: INFO nova.compute.manager [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Terminating instance [ 1329.422231] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]522e5429-c27e-525a-c617-251fb01fac13, 'name': SearchDatastore_Task, 'duration_secs': 0.013273} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.423109] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1329.423109] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Processing image 3ed5d656-a616-4609-8bb7-b02beb3ac3df {{(pid=65121) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1329.423109] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1329.424020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquired lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1329.424020] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1329.424020] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f44623cd-f208-428e-b922-825505f64266 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.434207] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65121) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1329.435624] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65121) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1329.435624] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c803ef51-eb09-4419-8af8-23c499aedf5b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.441819] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1329.441819] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]524a0341-7090-ad21-80e9-09dfcbba0ee0" [ 1329.441819] env[65121]: _type = "Task" [ 1329.441819] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.449090] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524a0341-7090-ad21-80e9-09dfcbba0ee0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.589252] env[65121]: DEBUG nova.compute.manager [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1329.589252] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1329.590385] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347bc7f4-9c3b-4b55-9357-9e0a64e4e471 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.601517] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1329.601811] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4da951a9-bd54-4010-9c22-b3c4d933ee02 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.610523] env[65121]: DEBUG oslo_vmware.api [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1329.610523] env[65121]: value = "task-5107695" [ 1329.610523] env[65121]: _type = "Task" [ 1329.610523] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.624265] env[65121]: DEBUG oslo_vmware.api [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.952516] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]524a0341-7090-ad21-80e9-09dfcbba0ee0, 'name': SearchDatastore_Task, 'duration_secs': 0.011221} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.953378] env[65121]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b707a76-6960-4199-8331-5ec075420bfd {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.959864] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1329.959864] env[65121]: value = "session[52394296-f348-19ca-7f3a-33db5ace4265]52de3973-ddbd-11c0-8da5-802916b3e77b" [ 1329.959864] env[65121]: _type = "Task" [ 1329.959864] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.967970] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52de3973-ddbd-11c0-8da5-802916b3e77b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.120997] env[65121]: DEBUG oslo_vmware.api [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107695, 'name': PowerOffVM_Task, 'duration_secs': 0.318819} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.121279] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1330.121442] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1330.121684] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-854fef9b-8c75-468f-bfb2-e3e1a7cc0cf0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.190276] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1330.190561] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1330.190811] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleting the datastore file [datastore1] d34b8457-adc1-4d6b-befd-f6f1aff6d1fb {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1330.191130] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d342f10-aa8d-40d8-9d61-c34443a51a26 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.199530] env[65121]: DEBUG oslo_vmware.api [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1330.199530] env[65121]: value = "task-5107697" [ 1330.199530] env[65121]: _type = "Task" [ 1330.199530] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.208207] env[65121]: DEBUG oslo_vmware.api [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107697, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.473932] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': session[52394296-f348-19ca-7f3a-33db5ace4265]52de3973-ddbd-11c0-8da5-802916b3e77b, 'name': SearchDatastore_Task, 'duration_secs': 0.010755} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.474264] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Releasing lock "[datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk" {{(pid=65121) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1330.474690] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 55d1d720-e0ba-4047-9820-43e6f4bbbc85/55d1d720-e0ba-4047-9820-43e6f4bbbc85.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1330.474861] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00bba2d9-ef39-4a64-9c1c-bba769854432 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.482852] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1330.482852] env[65121]: value = "task-5107698" [ 1330.482852] env[65121]: _type = "Task" [ 1330.482852] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.491635] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.711487] env[65121]: DEBUG oslo_vmware.api [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107697, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15306} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.711836] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1330.712082] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1330.712357] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1330.712678] env[65121]: INFO nova.compute.manager [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1330.712970] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1330.713202] env[65121]: DEBUG nova.compute.manager [-] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1330.713298] env[65121]: DEBUG nova.network.neutron [-] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1330.713669] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1330.714253] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1330.714595] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1330.763214] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1330.993653] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500493} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.993938] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/3ed5d656-a616-4609-8bb7-b02beb3ac3df/3ed5d656-a616-4609-8bb7-b02beb3ac3df.vmdk to [datastore1] 55d1d720-e0ba-4047-9820-43e6f4bbbc85/55d1d720-e0ba-4047-9820-43e6f4bbbc85.vmdk {{(pid=65121) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1330.994176] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Extending root virtual disk to 1048576 {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1330.994529] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-546b9835-bbdc-48ae-9af5-55f6ee84eed4 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.001720] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1331.001720] env[65121]: value = "task-5107699" [ 1331.001720] env[65121]: _type = "Task" [ 1331.001720] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.010977] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107699, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.220476] env[65121]: DEBUG nova.compute.manager [req-b9644987-de40-4c54-8372-3f144809caf0 req-6efbf900-c2b4-49a8-bafb-5ba8532ea968 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Received event network-vif-deleted-1c9a1903-516b-4056-ba71-cf778ceebf76 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1331.220943] env[65121]: INFO nova.compute.manager [req-b9644987-de40-4c54-8372-3f144809caf0 req-6efbf900-c2b4-49a8-bafb-5ba8532ea968 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Neutron deleted interface 1c9a1903-516b-4056-ba71-cf778ceebf76; detaching it from the instance and deleting it from the info cache [ 1331.221393] env[65121]: DEBUG nova.network.neutron [req-b9644987-de40-4c54-8372-3f144809caf0 req-6efbf900-c2b4-49a8-bafb-5ba8532ea968 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1331.512706] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107699, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.255639} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.513054] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Extended root virtual disk {{(pid=65121) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1331.513799] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a39a61c-f9a4-40b4-8ffa-9c94ddbb67e3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.536317] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Reconfiguring VM instance instance-0000007f to attach disk [datastore1] 55d1d720-e0ba-4047-9820-43e6f4bbbc85/55d1d720-e0ba-4047-9820-43e6f4bbbc85.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1331.536638] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-520f8432-7e41-49cd-a90f-af9b482b89f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.557886] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1331.557886] env[65121]: value = "task-5107700" [ 1331.557886] env[65121]: _type = "Task" [ 1331.557886] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.566441] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107700, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.707501] env[65121]: DEBUG nova.network.neutron [-] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1331.724211] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ca4b8b2-0d39-4be4-9b94-226a13d2e736 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.735591] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aac1e7-dfdc-426f-8b91-c05cf9049e9d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.768060] env[65121]: DEBUG nova.compute.manager [req-b9644987-de40-4c54-8372-3f144809caf0 req-6efbf900-c2b4-49a8-bafb-5ba8532ea968 service nova] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Detach interface failed, port_id=1c9a1903-516b-4056-ba71-cf778ceebf76, reason: Instance d34b8457-adc1-4d6b-befd-f6f1aff6d1fb could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1332.068632] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107700, 'name': ReconfigVM_Task, 'duration_secs': 0.312133} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.068766] env[65121]: DEBUG nova.virt.vmwareapi.volumeops [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Reconfigured VM instance instance-0000007f to attach disk [datastore1] 55d1d720-e0ba-4047-9820-43e6f4bbbc85/55d1d720-e0ba-4047-9820-43e6f4bbbc85.vmdk or device None with type sparse {{(pid=65121) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1332.069715] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e51fc7c-214d-42c5-8c3e-af661c02fd2b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.077370] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1332.077370] env[65121]: value = "task-5107701" [ 1332.077370] env[65121]: _type = "Task" [ 1332.077370] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.090344] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107701, 'name': Rename_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.212425] env[65121]: INFO nova.compute.manager [-] [instance: d34b8457-adc1-4d6b-befd-f6f1aff6d1fb] Took 1.50 seconds to deallocate network for instance. [ 1332.588233] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107701, 'name': Rename_Task, 'duration_secs': 0.146011} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.588685] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Powering on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1332.588732] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b1acbed-827e-4d66-8890-61cbc0bc69c8 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.595209] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1332.595209] env[65121]: value = "task-5107702" [ 1332.595209] env[65121]: _type = "Task" [ 1332.595209] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.603127] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107702, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.719136] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1332.719442] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1332.719768] env[65121]: DEBUG nova.objects.instance [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lazy-loading 'resources' on Instance uuid d34b8457-adc1-4d6b-befd-f6f1aff6d1fb {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1333.106126] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107702, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.289564] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a106d3f-3d43-40dd-9452-1d570c54febe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.298966] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861f1631-2049-4130-8d39-051a41b9f83d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.331199] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a630f65-4fa4-4baa-979f-f6c0f08b190d {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.339324] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bac54cc-bb2e-4eda-8b34-0930ddf925c3 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.355286] env[65121]: DEBUG nova.compute.provider_tree [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1333.608544] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107702, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.858693] env[65121]: DEBUG nova.scheduler.client.report [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1334.106979] env[65121]: DEBUG oslo_vmware.api [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107702, 'name': PowerOnVM_Task, 'duration_secs': 1.160773} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.107692] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Powered on the VM {{(pid=65121) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1334.107782] env[65121]: INFO nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Took 8.87 seconds to spawn the instance on the hypervisor. [ 1334.107954] env[65121]: DEBUG nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Checking state {{(pid=65121) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1334.108825] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36151a01-1687-46fc-a7a0-368142ba20f5 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.363701] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.644s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1334.389323] env[65121]: INFO nova.scheduler.client.report [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleted allocations for instance d34b8457-adc1-4d6b-befd-f6f1aff6d1fb [ 1334.627574] env[65121]: INFO nova.compute.manager [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Took 13.62 seconds to build instance. [ 1334.899787] env[65121]: DEBUG oslo_concurrency.lockutils [None req-a0ccb61e-be03-48f8-b568-332f2d848fa1 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d34b8457-adc1-4d6b-befd-f6f1aff6d1fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.819s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1335.129952] env[65121]: DEBUG oslo_concurrency.lockutils [None req-501d4324-18ac-48d8-bea4-59ecf8fc20f5 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.131s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1335.574153] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1335.574485] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1335.574742] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1335.575047] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1335.575047] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1335.577398] env[65121]: INFO nova.compute.manager [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Terminating instance [ 1336.081857] env[65121]: DEBUG nova.compute.manager [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1336.082220] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1336.083247] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f395b6e-24b7-4939-b112-23c1579bad0e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.091839] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1336.092134] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24be4815-c8d6-4023-af9e-3055055dd286 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.099174] env[65121]: DEBUG oslo_vmware.api [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1336.099174] env[65121]: value = "task-5107704" [ 1336.099174] env[65121]: _type = "Task" [ 1336.099174] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.109943] env[65121]: DEBUG oslo_vmware.api [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107704, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.609183] env[65121]: DEBUG oslo_vmware.api [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107704, 'name': PowerOffVM_Task, 'duration_secs': 0.222009} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.609441] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1336.609603] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1336.609854] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a54e4907-65d1-4730-bc17-ba0c24868254 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.678216] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1336.678492] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Deleting contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1336.678732] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Deleting the datastore file [datastore1] 55d1d720-e0ba-4047-9820-43e6f4bbbc85 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1336.679023] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-706a0a71-71c7-4de1-a74b-40f21c4a745f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.688970] env[65121]: DEBUG oslo_vmware.api [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for the task: (returnval){ [ 1336.688970] env[65121]: value = "task-5107706" [ 1336.688970] env[65121]: _type = "Task" [ 1336.688970] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.696955] env[65121]: DEBUG oslo_vmware.api [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107706, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.198698] env[65121]: DEBUG oslo_vmware.api [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Task: {'id': task-5107706, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14516} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.199113] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1337.199199] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Deleted contents of the VM from datastore datastore1 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1337.199314] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1337.199478] env[65121]: INFO nova.compute.manager [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1337.199714] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1337.199905] env[65121]: DEBUG nova.compute.manager [-] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1337.199996] env[65121]: DEBUG nova.network.neutron [-] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1337.200247] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1337.200797] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1337.201069] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1337.236931] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1337.490930] env[65121]: DEBUG nova.compute.manager [req-ca7258b9-797c-4ebf-ab0c-d3e2f23d0e3f req-30a9524c-3f5b-497c-bf9e-2e12bf653927 service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Received event network-vif-deleted-1f005be0-5cc6-421b-8b7f-769561de8d43 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1337.491012] env[65121]: INFO nova.compute.manager [req-ca7258b9-797c-4ebf-ab0c-d3e2f23d0e3f req-30a9524c-3f5b-497c-bf9e-2e12bf653927 service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Neutron deleted interface 1f005be0-5cc6-421b-8b7f-769561de8d43; detaching it from the instance and deleting it from the info cache [ 1337.491215] env[65121]: DEBUG nova.network.neutron [req-ca7258b9-797c-4ebf-ab0c-d3e2f23d0e3f req-30a9524c-3f5b-497c-bf9e-2e12bf653927 service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1337.967090] env[65121]: DEBUG nova.network.neutron [-] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1337.993613] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8c589f59-63bc-4675-8c53-e46071fab5ef {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.004825] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d8862f-d63a-4338-a998-cb5666742507 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.032261] env[65121]: DEBUG nova.compute.manager [req-ca7258b9-797c-4ebf-ab0c-d3e2f23d0e3f req-30a9524c-3f5b-497c-bf9e-2e12bf653927 service nova] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Detach interface failed, port_id=1f005be0-5cc6-421b-8b7f-769561de8d43, reason: Instance 55d1d720-e0ba-4047-9820-43e6f4bbbc85 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1338.469656] env[65121]: INFO nova.compute.manager [-] [instance: 55d1d720-e0ba-4047-9820-43e6f4bbbc85] Took 1.27 seconds to deallocate network for instance. [ 1338.976473] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1338.976823] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1338.977073] env[65121]: DEBUG nova.objects.instance [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lazy-loading 'resources' on Instance uuid 55d1d720-e0ba-4047-9820-43e6f4bbbc85 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1339.539238] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af446cb-76b3-47f1-a7f0-5be0ecda3304 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.547308] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411a6a91-7723-48b7-ad42-4f101c4f9f8e {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.579215] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262d39e5-6844-4e12-a03d-9c14de832e3a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.587418] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad51f0d-5fc9-4c1b-88e1-e4f826aa1d71 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.601527] env[65121]: DEBUG nova.compute.provider_tree [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1340.104766] env[65121]: DEBUG nova.scheduler.client.report [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1340.452924] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1340.453332] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1340.453613] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1340.453865] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1340.454097] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1340.457169] env[65121]: INFO nova.compute.manager [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Terminating instance [ 1340.610368] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1340.635401] env[65121]: INFO nova.scheduler.client.report [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Deleted allocations for instance 55d1d720-e0ba-4047-9820-43e6f4bbbc85 [ 1340.961402] env[65121]: DEBUG nova.compute.manager [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1340.961607] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1340.962520] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1c96ae-3321-428f-8911-d52bc2034b05 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.972188] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1340.972430] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffb9c460-c980-433a-b4e0-1fc4b0dc0ab7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.981382] env[65121]: DEBUG oslo_vmware.api [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1340.981382] env[65121]: value = "task-5107707" [ 1340.981382] env[65121]: _type = "Task" [ 1340.981382] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.989724] env[65121]: DEBUG oslo_vmware.api [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.143699] env[65121]: DEBUG oslo_concurrency.lockutils [None req-447cf77e-8a75-423d-964c-c2fb9a3d06b1 tempest-ServersNegativeTestMultiTenantJSON-105402919 tempest-ServersNegativeTestMultiTenantJSON-105402919-project-member] Lock "55d1d720-e0ba-4047-9820-43e6f4bbbc85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.569s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1341.492067] env[65121]: DEBUG oslo_vmware.api [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107707, 'name': PowerOffVM_Task, 'duration_secs': 0.213846} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.492286] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1341.492447] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1341.492702] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4628d43-088c-4a15-8d6a-bdd6e32684e7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.563220] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1341.563460] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1341.563645] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleting the datastore file [datastore2] d49a301c-c0d5-476e-9f1f-b8f98cf6adb7 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1341.563926] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-521f859f-82c1-4a24-9797-99741cdfa849 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.572125] env[65121]: DEBUG oslo_vmware.api [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1341.572125] env[65121]: value = "task-5107709" [ 1341.572125] env[65121]: _type = "Task" [ 1341.572125] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.580960] env[65121]: DEBUG oslo_vmware.api [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.082825] env[65121]: DEBUG oslo_vmware.api [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107709, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170544} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.083155] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1342.083333] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1342.083506] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1342.083673] env[65121]: INFO nova.compute.manager [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1342.083917] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1342.084128] env[65121]: DEBUG nova.compute.manager [-] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1342.084223] env[65121]: DEBUG nova.network.neutron [-] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1342.084462] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1342.084973] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1342.085250] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1342.129273] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1342.342170] env[65121]: DEBUG nova.compute.manager [req-46dc7b61-6954-436d-b371-5511f9ed8943 req-dd6f031d-2f54-40ef-ab95-e87747b10cdf service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Received event network-vif-deleted-10c1c247-c89e-4cd4-9ef3-0bf3621ad21c {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1342.342335] env[65121]: INFO nova.compute.manager [req-46dc7b61-6954-436d-b371-5511f9ed8943 req-dd6f031d-2f54-40ef-ab95-e87747b10cdf service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Neutron deleted interface 10c1c247-c89e-4cd4-9ef3-0bf3621ad21c; detaching it from the instance and deleting it from the info cache [ 1342.342446] env[65121]: DEBUG nova.network.neutron [req-46dc7b61-6954-436d-b371-5511f9ed8943 req-dd6f031d-2f54-40ef-ab95-e87747b10cdf service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1342.816446] env[65121]: DEBUG nova.network.neutron [-] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1342.845247] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b06f024-536b-4caf-b7b8-280304e7037b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.855998] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b44da2-1548-4afd-be3c-0fd29a7d8509 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.883644] env[65121]: DEBUG nova.compute.manager [req-46dc7b61-6954-436d-b371-5511f9ed8943 req-dd6f031d-2f54-40ef-ab95-e87747b10cdf service nova] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Detach interface failed, port_id=10c1c247-c89e-4cd4-9ef3-0bf3621ad21c, reason: Instance d49a301c-c0d5-476e-9f1f-b8f98cf6adb7 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1343.318716] env[65121]: INFO nova.compute.manager [-] [instance: d49a301c-c0d5-476e-9f1f-b8f98cf6adb7] Took 1.23 seconds to deallocate network for instance. [ 1343.824922] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1343.825252] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1343.825629] env[65121]: DEBUG nova.objects.instance [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lazy-loading 'resources' on Instance uuid d49a301c-c0d5-476e-9f1f-b8f98cf6adb7 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1344.377395] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e76667-eff1-432a-aa19-8353d32f32be {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.385890] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7117bd08-adf2-4b84-92a8-54823df93bde {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.419243] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfbb8be-80f9-4530-adf3-4cb2ff2c2060 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.427406] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3311be-af16-4b11-9e81-b984308e8296 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.443767] env[65121]: DEBUG nova.compute.provider_tree [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1344.947794] env[65121]: DEBUG nova.scheduler.client.report [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1345.454139] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.628s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1345.475692] env[65121]: INFO nova.scheduler.client.report [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleted allocations for instance d49a301c-c0d5-476e-9f1f-b8f98cf6adb7 [ 1345.984697] env[65121]: DEBUG oslo_concurrency.lockutils [None req-d9d39edd-29d2-4d8d-832c-186b0e939eda tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "d49a301c-c0d5-476e-9f1f-b8f98cf6adb7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.531s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1347.160584] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1347.161056] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1347.161204] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "3859dd02-9b52-4c71-a41f-c41f47ac1706-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1347.161925] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1347.162177] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1347.164912] env[65121]: INFO nova.compute.manager [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Terminating instance [ 1347.668841] env[65121]: DEBUG nova.compute.manager [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Start destroying the instance on the hypervisor. {{(pid=65121) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1347.669115] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Destroying instance {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1347.670062] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1875d6-0c74-4b08-b1a1-e1434a619d59 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.678305] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powering off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1347.678579] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0686a8f0-7bd3-4dec-86f9-e0fa0e5b5f9f {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.686500] env[65121]: DEBUG oslo_vmware.api [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1347.686500] env[65121]: value = "task-5107710" [ 1347.686500] env[65121]: _type = "Task" [ 1347.686500] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.695927] env[65121]: DEBUG oslo_vmware.api [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107710, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.197462] env[65121]: DEBUG oslo_vmware.api [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107710, 'name': PowerOffVM_Task, 'duration_secs': 0.275376} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.197462] env[65121]: DEBUG nova.virt.vmwareapi.vm_util [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Powered off the VM {{(pid=65121) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1348.198146] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Unregistering the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1348.198146] env[65121]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d91158f0-bf6d-423a-81eb-c5d0957b378b {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.270893] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Unregistered the VM {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1348.271246] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Deleting contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1348.271451] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleting the datastore file [datastore2] 3859dd02-9b52-4c71-a41f-c41f47ac1706 {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1348.271743] env[65121]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16f0c42e-9b03-4828-8202-df0f68f1eeec {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.279191] env[65121]: DEBUG oslo_vmware.api [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for the task: (returnval){ [ 1348.279191] env[65121]: value = "task-5107712" [ 1348.279191] env[65121]: _type = "Task" [ 1348.279191] env[65121]: } to complete. {{(pid=65121) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.288107] env[65121]: DEBUG oslo_vmware.api [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107712, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.789904] env[65121]: DEBUG oslo_vmware.api [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Task: {'id': task-5107712, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150161} completed successfully. {{(pid=65121) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.790301] env[65121]: DEBUG nova.virt.vmwareapi.ds_util [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleted the datastore file {{(pid=65121) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1348.790548] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Deleted contents of the VM from datastore datastore2 {{(pid=65121) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1348.790820] env[65121]: DEBUG nova.virt.vmwareapi.vmops [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Instance destroyed {{(pid=65121) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1348.790936] env[65121]: INFO nova.compute.manager [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1348.791212] env[65121]: DEBUG oslo.service.backend._common.loopingcall [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65121) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1348.791405] env[65121]: DEBUG nova.compute.manager [-] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Deallocating network for instance {{(pid=65121) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1348.791497] env[65121]: DEBUG nova.network.neutron [-] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] deallocate_for_instance() {{(pid=65121) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1348.791785] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1348.792339] env[65121]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1348.792594] env[65121]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1348.830109] env[65121]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1349.091683] env[65121]: DEBUG nova.compute.manager [req-0ab54559-49e5-481e-a311-538adad1c2bf req-cb90ba8f-d8e1-4839-baa3-c06275fdfa04 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Received event network-vif-deleted-f714bfb3-87de-4c6b-ba80-8227e6eebf65 {{(pid=65121) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1349.091923] env[65121]: INFO nova.compute.manager [req-0ab54559-49e5-481e-a311-538adad1c2bf req-cb90ba8f-d8e1-4839-baa3-c06275fdfa04 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Neutron deleted interface f714bfb3-87de-4c6b-ba80-8227e6eebf65; detaching it from the instance and deleting it from the info cache [ 1349.092111] env[65121]: DEBUG nova.network.neutron [req-0ab54559-49e5-481e-a311-538adad1c2bf req-cb90ba8f-d8e1-4839-baa3-c06275fdfa04 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1349.565747] env[65121]: DEBUG nova.network.neutron [-] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Updating instance_info_cache with network_info: [] {{(pid=65121) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1349.594536] env[65121]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63671729-5b29-4481-8e4f-06fa08c34db2 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.605222] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b869af55-4e58-4508-b482-7b91602de35a {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.633765] env[65121]: DEBUG nova.compute.manager [req-0ab54559-49e5-481e-a311-538adad1c2bf req-cb90ba8f-d8e1-4839-baa3-c06275fdfa04 service nova] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Detach interface failed, port_id=f714bfb3-87de-4c6b-ba80-8227e6eebf65, reason: Instance 3859dd02-9b52-4c71-a41f-c41f47ac1706 could not be found. {{(pid=65121) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1350.068668] env[65121]: INFO nova.compute.manager [-] [instance: 3859dd02-9b52-4c71-a41f-c41f47ac1706] Took 1.28 seconds to deallocate network for instance. [ 1350.575619] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1350.576093] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1350.576230] env[65121]: DEBUG nova.objects.instance [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lazy-loading 'resources' on Instance uuid 3859dd02-9b52-4c71-a41f-c41f47ac1706 {{(pid=65121) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1351.115881] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c795eb32-7116-46b3-97ac-7db33d2a2301 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.123822] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213ec773-1273-4471-9618-dd64e4d40911 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.155054] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7cdee7-bfaa-4f8f-8cf3-caa083510a91 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.163186] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba58b2c-b77d-421f-8403-4c2160ade9f7 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.177618] env[65121]: DEBUG nova.compute.provider_tree [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1351.681063] env[65121]: DEBUG nova.scheduler.client.report [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1351.912212] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.912460] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.912614] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.912757] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.912906] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.913092] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.913254] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.913391] env[65121]: DEBUG nova.compute.manager [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65121) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1351.913532] env[65121]: DEBUG oslo_service.periodic_task [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65121) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.186372] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1352.205354] env[65121]: INFO nova.scheduler.client.report [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Deleted allocations for instance 3859dd02-9b52-4c71-a41f-c41f47ac1706 [ 1352.416405] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1352.416678] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1352.416816] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1352.416968] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65121) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1352.417911] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e5abc1-964c-4a06-b05c-85a5c406c764 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.426629] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b83ce9-b486-4440-9686-c5178d929d8c {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.440925] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0f29a3-790a-4d0b-9d89-2d69fa527883 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.448057] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc2cefa-a043-4a7e-a5cc-c61b1f02c788 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.477946] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178827MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65121) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1352.478192] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1352.478332] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1352.713020] env[65121]: DEBUG oslo_concurrency.lockutils [None req-edeca43b-7977-43bd-b629-017fcc1e3fa4 tempest-ServerRescueNegativeTestJSON-1990305895 tempest-ServerRescueNegativeTestJSON-1990305895-project-member] Lock "3859dd02-9b52-4c71-a41f-c41f47ac1706" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.552s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1353.497722] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1353.497925] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65121) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1353.512071] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3006fb-37d6-46fc-81a0-924c291a0008 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.521012] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a4a493-1de1-40a5-a1bb-4ab9843bcefe {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.553199] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2847ebd-da54-4b40-9ab1-877ab0251ef0 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.561303] env[65121]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933e7b23-5280-4874-8b40-666d5a31cb01 {{(pid=65121) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.575294] env[65121]: DEBUG nova.compute.provider_tree [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed in ProviderTree for provider: d56783bf-3ede-475a-8c5a-8d8303049e47 {{(pid=65121) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1354.078619] env[65121]: DEBUG nova.scheduler.client.report [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Inventory has not changed for provider d56783bf-3ede-475a-8c5a-8d8303049e47 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65121) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1354.584058] env[65121]: DEBUG nova.compute.resource_tracker [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65121) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1354.584251] env[65121]: DEBUG oslo_concurrency.lockutils [None req-3cc7e2d6-aee7-488f-a69f-fcd692e21824 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.106s {{(pid=65121) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}}